Controlling aliased dynamics in motion systems? An identification for sampled-data control approach
NASA Astrophysics Data System (ADS)
Oomen, Tom
2014-07-01
Sampled-data control systems occasionally exhibit aliased resonance phenomena within the control bandwidth. The aim of this paper is to investigate the aspect of these aliased dynamics with application to a high performance industrial nano-positioning machine. This necessitates a full sampled-data control design approach, since these aliased dynamics endanger both the at-sample performance and the intersample behaviour. The proposed framework comprises both system identification and sampled-data control. In particular, the sampled-data control objective necessitates models that encompass the intersample behaviour, i.e., ideally continuous time models. Application of the proposed approach on an industrial wafer stage system provides a thorough insight and new control design guidelines for controlling aliased dynamics.
Stochastic Stability of Sampled Data Systems with a Jump Linear Controller
NASA Technical Reports Server (NTRS)
Gonzalez, Oscar R.; Herencia-Zapana, Heber; Gray, W. Steven
2004-01-01
In this paper an equivalence between the stochastic stability of a sampled-data system and its associated discrete-time representation is established. The sampled-data system consists of a deterministic, linear, time-invariant, continuous-time plant and a stochastic, linear, time-invariant, discrete-time, jump linear controller. The jump linear controller models computer systems and communication networks that are subject to stochastic upsets or disruptions. This sampled-data model has been used in the analysis and design of fault-tolerant systems and computer-control systems with random communication delays without taking into account the inter-sample response. This paper shows that the known equivalence between the stability of a deterministic sampled-data system and the associated discrete-time representation holds even in a stochastic framework.
Stochastic Stability of Nonlinear Sampled Data Systems with a Jump Linear Controller
NASA Technical Reports Server (NTRS)
Gonzalez, Oscar R.; Herencia-Zapana, Heber; Gray, W. Steven
2004-01-01
This paper analyzes the stability of a sampled- data system consisting of a deterministic, nonlinear, time- invariant, continuous-time plant and a stochastic, discrete- time, jump linear controller. The jump linear controller mod- els, for example, computer systems and communication net- works that are subject to stochastic upsets or disruptions. This sampled-data model has been used in the analysis and design of fault-tolerant systems and computer-control systems with random communication delays without taking into account the inter-sample response. To analyze stability, appropriate topologies are introduced for the signal spaces of the sampled- data system. With these topologies, the ideal sampling and zero-order-hold operators are shown to be measurable maps. This paper shows that the known equivalence between the stability of a deterministic, linear sampled-data system and its associated discrete-time representation as well as between a nonlinear sampled-data system and a linearized representation holds even in a stochastic framework.
Systems and methods for self-synchronized digital sampling
NASA Technical Reports Server (NTRS)
Samson, Jr., John R. (Inventor)
2008-01-01
Systems and methods for self-synchronized data sampling are provided. In one embodiment, a system for capturing synchronous data samples is provided. The system includes an analog to digital converter adapted to capture signals from one or more sensors and convert the signals into a stream of digital data samples at a sampling frequency determined by a sampling control signal; and a synchronizer coupled to the analog to digital converter and adapted to receive a rotational frequency signal from a rotating machine, wherein the synchronizer is further adapted to generate the sampling control signal, and wherein the sampling control signal is based on the rotational frequency signal.
Modular Biometric Monitoring System
NASA Technical Reports Server (NTRS)
Chmiel, Alan J. (Inventor); Humphreys, Bradley T. (Inventor)
2017-01-01
A modular system for acquiring biometric data includes a plurality of data acquisition modules configured to sample biometric data from at least one respective input channel at a data acquisition rate. A representation of the sampled biometric data is stored in memory of each of the plurality of data acquisition modules. A central control system is in communication with each of the plurality of data acquisition modules through a bus. The central control system is configured to control communication of data, via the bus, with each of the plurality of data acquisition modules.
NASA Technical Reports Server (NTRS)
Kuo, B. C.; Singh, G.
1974-01-01
The dynamics of the Large Space Telescope (LST) control system were studied in order to arrive at a simplified model for computer simulation without loss of accuracy. The frictional nonlinearity of the Control Moment Gyroscope (CMG) Control Loop was analyzed in a model to obtain data for the following: (1) a continuous describing function for the gimbal friction nonlinearity; (2) a describing function of the CMG nonlinearity using an analytical torque equation; and (3) the discrete describing function and function plots for CMG functional linearity. Preliminary computer simulations are shown for the simplified LST system, first without, and then with analytical torque expressions. Transfer functions of the sampled-data LST system are also described. A final computer simulation is presented which uses elements of the simplified sampled-data LST system with analytical CMG frictional torque expressions.
Digital flight control systems
NASA Technical Reports Server (NTRS)
Caglayan, A. K.; Vanlandingham, H. F.
1977-01-01
The design of stable feedback control laws for sampled-data systems with variable rate sampling was investigated. These types of sampled-data systems arise naturally in digital flight control systems which use digital actuators where it is desirable to decrease the number of control computer output commands in order to save wear and tear of the associated equipment. The design of aircraft control systems which are optimally tolerant of sensor and actuator failures was also studied. Detection of the failed sensor or actuator must be resolved and if the estimate of the state is used in the control law, then it is also desirable to have an estimator which will give the optimal state estimate even under the failed conditions.
NASA Astrophysics Data System (ADS)
Zhu, Baolong; Zhang, Zhiping; Zhou, Ding; Ma, Jie; Li, Shunli
2017-08-01
This paper investigates the H∞ control problem of the attitude stabilisation of a rigid spacecraft with external disturbances using prediction-based sampled-data control strategy. Aiming to achieve a 'virtual' closed-loop system, a type of parameterised sampled-data controller is designed by introducing a prediction mechanism. The resultant closed-loop system is equivalent to a hybrid system featured by a continuous-time and an impulsive differential system. By using a time-varying Lyapunov functional, a generalised bounded real lemma (GBRL) is first established for a kind of impulsive differential system. Based on this GBRL and Lyapunov functional approach, a sufficient condition is derived to guarantee the closed-loop system to be asymptotically stable and to achieve a prescribed H∞ performance. In addition, the controller parameter tuning is cast into a convex optimisation problem. Simulation and comparative results are provided to illustrate the effectiveness of the developed control scheme.
Stabilization for sampled-data neural-network-based control systems.
Zhu, Xun-Lin; Wang, Youyi
2011-02-01
This paper studies the problem of stabilization for sampled-data neural-network-based control systems with an optimal guaranteed cost. Unlike previous works, the resulting closed-loop system with variable uncertain sampling cannot simply be regarded as an ordinary continuous-time system with a fast-varying delay in the state. By defining a novel piecewise Lyapunov functional and using a convex combination technique, the characteristic of sampled-data systems is captured. A new delay-dependent stabilization criterion is established in terms of linear matrix inequalities such that the maximal sampling interval and the minimal guaranteed cost control performance can be obtained. It is shown that the newly proposed approach can lead to less conservative and less complex results than the existing ones. Application examples are given to illustrate the effectiveness and the benefits of the proposed method.
NASA Astrophysics Data System (ADS)
Ji, Yinghua; Ju-Ju, Hu; Jian-Hua, Huang; Qiang, Ke
Due to the influence of decoherence, the quantum state probably evolves from the initial pure state to the mixed state, resulting in loss of fidelity, coherence and purity, which is deteriorating for quantum information transmission. Thus, in quantum engineering, quantum control should not only realize the transfer and track of quantum states through manipulation of the external electromagnetic field but also enhance the robustness against decoherence. In this paper, we aim to design a control law to steer the system into the sliding mode domain and maintain it in that domain when bounded uncertainties exist in the system Hamiltonian. We first define the required control performance by fidelity, degree of coherence and purity in terms of the uncertainty of the Hamiltonian in Markovian open quantum system. By characterizing the required robustness using a sliding mode domain, a sampled-data design method is introduced for decoherence control in the quantum system. Furthermore, utilizing the sampled data, a control scheme has been designed on the basis of sliding mode control, and the choice of sampling operator and driving of quantum state during the sampling by the Lyapunov control method are discussed.
Lam, H K
2012-02-01
This paper investigates the stability of sampled-data output-feedback (SDOF) polynomial-fuzzy-model-based control systems. Representing the nonlinear plant using a polynomial fuzzy model, an SDOF fuzzy controller is proposed to perform the control process using the system output information. As only the system output is available for feedback compensation, it is more challenging for the controller design and system analysis compared to the full-state-feedback case. Furthermore, because of the sampling activity, the control signal is kept constant by the zero-order hold during the sampling period, which complicates the system dynamics and makes the stability analysis more difficult. In this paper, two cases of SDOF fuzzy controllers, which either share the same number of fuzzy rules or not, are considered. The system stability is investigated based on the Lyapunov stability theory using the sum-of-squares (SOS) approach. SOS-based stability conditions are obtained to guarantee the system stability and synthesize the SDOF fuzzy controller. Simulation examples are given to demonstrate the merits of the proposed SDOF fuzzy control approach.
Multirate sampled-data yaw-damper and modal suppression system design
NASA Technical Reports Server (NTRS)
Berg, Martin C.; Mason, Gregory S.
1990-01-01
A multirate control law synthesized algorithm based on an infinite-time quadratic cost function, was developed along with a method for analyzing the robustness of multirate systems. A generalized multirate sampled-data control law structure (GMCLS) was introduced. A new infinite-time-based parameter optimization multirate sampled-data control law synthesis method and solution algorithm were developed. A singular-value-based method for determining gain and phase margins for multirate systems was also developed. The finite-time-based parameter optimization multirate sampled-data control law synthesis algorithm originally intended to be applied to the aircraft problem was instead demonstrated by application to a simpler problem involving the control of the tip position of a two-link robot arm. The GMCLS, the infinite-time-based parameter optimization multirate control law synthesis method and solution algorithm, and the singular-value based method for determining gain and phase margins were all demonstrated by application to the aircraft control problem originally proposed for this project.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
Covered are: analytical laboratory operations (ALO) sample receipt and control, ALO data report/package preparation review and control, single shell tank (PST) project sample tracking system, sample receiving, analytical balances, duties and responsibilities of sample custodian, sample refrigerator temperature monitoring, security, assignment of staff responsibilities, sample storage, data reporting, and general requirements for glassware.
Progress in multirate digital control system design
NASA Technical Reports Server (NTRS)
Berg, Martin C.; Mason, Gregory S.
1991-01-01
A new methodology for multirate sampled-data control design based on a new generalized control law structure, two new parameter-optimization-based control law synthesis methods, and a new singular-value-based robustness analysis method are described. The control law structure can represent multirate sampled-data control laws of arbitrary structure and dynamic order, with arbitrarily prescribed sampling rates for all sensors and update rates for all processor states and actuators. The two control law synthesis methods employ numerical optimization to determine values for the control law parameters. The robustness analysis method is based on the multivariable Nyquist criterion applied to the loop transfer function for the sampling period equal to the period of repetition of the system's complete sampling/update schedule. The complete methodology is demonstrated by application to the design of a combination yaw damper and modal suppression system for a commercial aircraft.
A passivity criterion for sampled-data bilateral teleoperation systems.
Jazayeri, Ali; Tavakoli, Mahdi
2013-01-01
A teleoperation system consists of a teleoperator, a human operator, and a remote environment. Conditions involving system and controller parameters that ensure the teleoperator passivity can serve as control design guidelines to attain maximum teleoperation transparency while maintaining system stability. In this paper, sufficient conditions for teleoperator passivity are derived for when position error-based controllers are implemented in discrete-time. This new analysis is necessary because discretization causes energy leaks and does not necessarily preserve the passivity of the system. The proposed criterion for sampled-data teleoperator passivity imposes lower bounds on the teleoperator's robots dampings, an upper bound on the sampling time, and bounds on the control gains. The criterion is verified through simulations and experiments.
Robust stability bounds for multi-delay networked control systems
NASA Astrophysics Data System (ADS)
Seitz, Timothy; Yedavalli, Rama K.; Behbahani, Alireza
2018-04-01
In this paper, the robust stability of a perturbed linear continuous-time system is examined when controlled using a sampled-data networked control system (NCS) framework. Three new robust stability bounds on the time-invariant perturbations to the original continuous-time plant matrix are presented guaranteeing stability for the corresponding discrete closed-loop augmented delay-free system (ADFS) with multiple time-varying sensor and actuator delays. The bounds are differentiated from previous work by accounting for the sampled-data nature of the NCS and for separate communication delays for each sensor and actuator, not a single delay. Therefore, this paper expands the knowledge base in multiple inputs multiple outputs (MIMO) sampled-data time delay systems. Bounds are presented for unstructured, semi-structured, and structured perturbations.
Data Acquisition for Modular Biometric Monitoring System
NASA Technical Reports Server (NTRS)
Grodsinsky, Carlos M. (Inventor); Chmiel, Alan J. (Inventor); Humphreys, Bradley T. (Inventor)
2014-01-01
A modular system for acquiring biometric data includes a plurality of data acquisition modules configured to sample biometric data from at least one respective input channel at a data acquisition rate. A representation of the sampled biometric data is stored in memory of each of the plurality of data acquisition modules. A central control system is in communication with each of the plurality of data acquisition modules through a bus. The central control system is configured to collect data asynchronously, via the bus, from the memory of the plurality of data acquisition modules according to a relative fullness of the memory of the plurality of data acquisition modules.
Vibration Pattern Imager (VPI): A control and data acquisition system for scanning laser vibrometers
NASA Technical Reports Server (NTRS)
Rizzi, Stephen A.; Brown, Donald E.; Shaffer, Thomas A.
1993-01-01
The Vibration Pattern Imager (VPI) system was designed to control and acquire data from scanning laser vibrometer sensors. The PC computer based system uses a digital signal processing (DSP) board and an analog I/O board to control the sensor and to process the data. The VPI system was originally developed for use with the Ometron VPI Sensor, but can be readily adapted to any commercially available sensor which provides an analog output signal and requires analog inputs for control of mirror positioning. The sensor itself is not part of the VPI system. A graphical interface program, which runs on a PC under the MS-DOS operating system, functions in an interactive mode and communicates with the DSP and I/O boards in a user-friendly fashion through the aid of pop-up menus. Two types of data may be acquired with the VPI system: single point or 'full field.' In the single point mode, time series data is sampled by the A/D converter on the I/O board (at a user-defined sampling rate for a selectable number of samples) and is stored by the PC. The position of the measuring point (adjusted by mirrors in the sensor) is controlled via a mouse input. The mouse input is translated to output voltages by the D/A converter on the I/O board to control the mirror servos. In the 'full field' mode, the measurement point is moved over a user-selectable rectangular area. The time series data is sampled by the A/D converter on the I/O board (at a user-defined sampling rate for a selectable number of samples) and converted to a root-mean-square (rms) value by the DSP board. The rms 'full field' velocity distribution is then uploaded for display and storage on the PC.
ERIC Educational Resources Information Center
Mitchell, Eugene E., Ed.
The simulation of a sampled-data system is described that uses a full parallel hybrid computer. The sampled data system simulated illustrates the proportional-integral-derivative (PID) discrete control of a continuous second-order process representing a stirred-tank. The stirred-tank is simulated using continuous analog components, while PID…
Improved synchronization criteria for time-delayed chaotic Lur'e systems using sampled-data control
NASA Astrophysics Data System (ADS)
Duan, Wenyong; Li, Yan; Fu, Xiaorong; Du, Baozhu
2017-02-01
This paper is concerned with the synchronization for a class of time-delayed chaotic Lur’e systems using sampled-data control. Both of time-varying and time-invariant delays are considered. New criteria are proposed in terms of linear matrix inequalities (LMIs) by employing a modified LKF combined with the delay-fraction theory and some novel terms. The criteria are less conservative than some previous ones and a longer sampling period is achieved under the new results. Furthermore, the derived conditions are employed to design a sampled-data controller. The desired controller gain matrix can be obtained by means of the LMI approach. Finally, a numerical examples and simulations on Chua’s circuit is presented to show the effectiveness of the proposed approach.
Networked event-triggered control: an introduction and research trends
NASA Astrophysics Data System (ADS)
Mahmoud, Magdi S.; Sabih, Muhammad
2014-11-01
A physical system can be studied as either continuous time or discrete-time system depending upon the control objectives. Discrete-time control systems can be further classified into two categories based on the sampling: (1) time-triggered control systems and (2) event-triggered control systems. Time-triggered systems sample states and calculate controls at every sampling instant in a periodic fashion, even in cases when states and calculated control do not change much. This indicates unnecessary and useless data transmission and computation efforts of a time-triggered system, thus inefficiency. For networked systems, the transmission of measurement and control signals, thus, cause unnecessary network traffic. Event-triggered systems, on the other hand, have potential to reduce the communication burden in addition to reducing the computation of control signals. This paper provides an up-to-date survey on the event-triggered methods for control systems and highlights the potential research directions.
Research study on stabilization and control: Modern sampled data control theory
NASA Technical Reports Server (NTRS)
Kuo, B. C.; Singh, G.; Yackel, R. A.
1973-01-01
A numerical analysis of spacecraft stability parameters was conducted. The analysis is based on a digital approximation by point by point state comparison. The technique used is that of approximating a continuous data system by a sampled data model by comparison of the states of the two systems. Application of the method to the digital redesign of the simplified one axis dynamics of the Skylab is presented.
Modernization of B-2 Data, Video, and Control Systems Infrastructure
NASA Technical Reports Server (NTRS)
Cmar, Mark D.; Maloney, Christian T.; Butala, Vishal D.
2012-01-01
The National Aeronautics and Space Administration (NASA) Glenn Research Center (GRC) Plum Brook Station (PBS) Spacecraft Propulsion Research Facility, commonly referred to as B-2, is NASA s third largest thermal-vacuum facility with propellant systems capability. B-2 has completed a modernization effort of its facility legacy data, video and control systems infrastructure to accommodate modern integrated testing and Information Technology (IT) Security requirements. Integrated systems tests have been conducted to demonstrate the new data, video and control systems functionality and capability. Discrete analog signal conditioners have been replaced by new programmable, signal processing hardware that is integrated with the data system. This integration supports automated calibration and verification of the analog subsystem. Modern measurement systems analysis (MSA) tools are being developed to help verify system health and measurement integrity. Legacy hard wired digital data systems have been replaced by distributed Fibre Channel (FC) network connected digitizers where high speed sampling rates have increased to 256,000 samples per second. Several analog video cameras have been replaced by digital image and storage systems. Hard-wired analog control systems have been replaced by Programmable Logic Controllers (PLC), fiber optic networks (FON) infrastructure and human machine interface (HMI) operator screens. New modern IT Security procedures and schemes have been employed to control data access and process control flows. Due to the nature of testing possible at B-2, flexibility and configurability of systems has been central to the architecture during modernization.
Modernization of B-2 Data, Video, and Control Systems Infrastructure
NASA Technical Reports Server (NTRS)
Cmar, Mark D.; Maloney, Christian T.; Butala, Vishal D.
2012-01-01
The National Aeronautics and Space Administration (NASA) Glenn Research Center (GRC) Plum Brook Station (PBS) Spacecraft Propulsion Research Facility, commonly referred to as B-2, is NASA's third largest thermal-vacuum facility with propellant systems capability. B-2 has completed a modernization effort of its facility legacy data, video and control systems infrastructure to accommodate modern integrated testing and Information Technology (IT) Security requirements. Integrated systems tests have been conducted to demonstrate the new data, video and control systems functionality and capability. Discrete analog signal conditioners have been replaced by new programmable, signal processing hardware that is integrated with the data system. This integration supports automated calibration and verification of the analog subsystem. Modern measurement systems analysis (MSA) tools are being developed to help verify system health and measurement integrity. Legacy hard wired digital data systems have been replaced by distributed Fibre Channel (FC) network connected digitizers where high speed sampling rates have increased to 256,000 samples per second. Several analog video cameras have been replaced by digital image and storage systems. Hard-wired analog control systems have been replaced by Programmable Logic Controllers (PLC), fiber optic networks (FON) infrastructure and human machine interface (HMI) operator screens. New modern IT Security procedures and schemes have been employed to control data access and process control flows. Due to the nature of testing possible at B-2, flexibility and configurability of systems has been central to the architecture during modernization.
An error criterion for determining sampling rates in closed-loop control systems
NASA Technical Reports Server (NTRS)
Brecher, S. M.
1972-01-01
The determination of an error criterion which will give a sampling rate for adequate performance of linear, time-invariant closed-loop, discrete-data control systems was studied. The proper modelling of the closed-loop control system for characterization of the error behavior, and the determination of an absolute error definition for performance of the two commonly used holding devices are discussed. The definition of an adequate relative error criterion as a function of the sampling rate and the parameters characterizing the system is established along with the determination of sampling rates. The validity of the expressions for the sampling interval was confirmed by computer simulations. Their application solves the problem of making a first choice in the selection of sampling rates.
Zeng, Cheng; Liang, Shan; Xiang, Shuwen
2017-05-01
Continuous-time systems are usually modelled by the form of ordinary differential equations arising from physical laws. However, the use of these models in practice and utilizing, analyzing or transmitting these data from such systems must first invariably be discretized. More importantly, for digital control of a continuous-time nonlinear system, a good sampled-data model is required. This paper investigates the new consistency condition which is weaker than the previous similar results presented. Moreover, given the stability of the high-order approximate model with stable zero dynamics, the novel condition presented stabilizes the exact sampled-data model of the nonlinear system for sufficiently small sampling periods. An insightful interpretation of the obtained results can be made in terms of the stable sampling zero dynamics, and the new consistency condition is surprisingly associated with the relative degree of the nonlinear continuous-time system. Our controller design, based on the higher-order approximate discretized model, extends the existing methods which mainly deal with the Euler approximation. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.
BROADBAND DIGITAL GEOPHYSICAL TELEMETRY SYSTEM.
Seeley, Robert L.; Daniels, Jeffrey J.
1984-01-01
A system has been developed to simultaneously sample and transmit digital data from five remote geophysical data receiver stations to a control station that processes, displays, and stores the data. A microprocessor in each remote station receives commands from the control station over a single telemetry channel.
NASA Technical Reports Server (NTRS)
Wilson, J. L.
1974-01-01
A users guide to the Sampled Data Stability Analysis Program (SADSAP) is provided. This program is a general purpose sampled data Stability Analysis Program capable of providing frequency response on root locus data.
Sample Manipulation System for Sample Analysis at Mars
NASA Technical Reports Server (NTRS)
Mumm, Erik; Kennedy, Tom; Carlson, Lee; Roberts, Dustyn
2008-01-01
The Sample Analysis at Mars (SAM) instrument will analyze Martian samples collected by the Mars Science Laboratory Rover with a suite of spectrometers. This paper discusses the driving requirements, design, and lessons learned in the development of the Sample Manipulation System (SMS) within SAM. The SMS stores and manipulates 74 sample cups to be used for solid sample pyrolysis experiments. Focus is given to the unique mechanism architecture developed to deliver a high packing density of sample cups in a reliable, fault tolerant manner while minimizing system mass and control complexity. Lessons learned are presented on contamination control, launch restraint mechanisms for fragile sample cups, and mechanism test data.
Computer graphics for quality control in the INAA of geological samples
Grossman, J.N.; Baedecker, P.A.
1987-01-01
A data reduction system for the routine instrumental activation analysis of samples is described, with particular emphasis on interactive graphics capabilities for evaluating analytical quality. Graphics procedures have been developed to interactively control the analysis of selected photopeaks during spectral analysis, and to evaluate detector performance during a given counting cycle. Graphics algorithms are also used to compare the data on reference samples with accepted values, to prepare quality control charts to evaluate long term precision and to search for systematic variations in data on reference samples as a function of time. ?? 1987 Akade??miai Kiado??.
Data processing for water monitoring system
NASA Technical Reports Server (NTRS)
Monford, L.; Linton, A. T.
1978-01-01
Water monitoring data acquisition system is structured about central computer that controls sampling and sensor operation, and analyzes and displays data in real time. Unit is essentially separated into two systems: computer system, and hard wire backup system which may function separately or with computer.
Local synchronization of chaotic neural networks with sampled-data and saturating actuators.
Wu, Zheng-Guang; Shi, Peng; Su, Hongye; Chu, Jian
2014-12-01
This paper investigates the problem of local synchronization of chaotic neural networks with sampled-data and actuator saturation. A new time-dependent Lyapunov functional is proposed for the synchronization error systems. The advantage of the constructed Lyapunov functional lies in the fact that it is positive definite at sampling times but not necessarily between sampling times, and makes full use of the available information about the actual sampling pattern. A local stability condition of the synchronization error systems is derived, based on which a sampled-data controller with respect to the actuator saturation is designed to ensure that the master neural networks and slave neural networks are locally asymptotically synchronous. Two optimization problems are provided to compute the desired sampled-data controller with the aim of enlarging the set of admissible initial conditions or the admissible sampling upper bound ensuring the local synchronization of the considered chaotic neural networks. A numerical example is used to demonstrate the effectiveness of the proposed design technique.
Pulse-Flow Microencapsulation System
NASA Technical Reports Server (NTRS)
Morrison, Dennis R.
2006-01-01
The pulse-flow microencapsulation system (PFMS) is an automated system that continuously produces a stream of liquid-filled microcapsules for delivery of therapeutic agents to target tissues. Prior microencapsulation systems have relied on batch processes that involve transfer of batches between different apparatuses for different stages of production followed by sampling for acquisition of quality-control data, including measurements of size. In contrast, the PFMS is a single, microprocessor-controlled system that performs all processing steps, including acquisition of quality-control data. The quality-control data can be used as real-time feedback to ensure the production of large quantities of uniform microcapsules.
An automated atmospheric sampling system operating on 747 airliners
NASA Technical Reports Server (NTRS)
Perkins, P. J.; Gustafsson, U. R. C.
1976-01-01
An air sampling system that automatically measures the temporal and spatial distribution of particulate and gaseous constituents of the atmosphere is collecting data on commercial air routes covering the world. Measurements are made in the upper troposphere and lower stratosphere (6 to 12 km) of constituents related to aircraft engine emissions and other pollutants. Aircraft operated by different airlines sample air at latitudes from the Arctic to Australia. This unique system includes specialized instrumentation, a special air inlet probe for sampling outside air, a computerized automatic control, and a data acquisition system. Air constituent and related flight data are tape recorded in flight for later computer processing on the ground.
An automated system for global atmospheric sampling using B-747 airliners
NASA Technical Reports Server (NTRS)
Lew, K. Q.; Gustafsson, U. R. C.; Johnson, R. E.
1981-01-01
The global air sampling program utilizes commercial aircrafts in scheduled service to measure atmospheric constituents. A fully automated system designed for the 747 aircraft is described. Airline operational constraints and data and control subsystems are treated. The overall program management, system monitoring, and data retrieval from four aircraft in global service is described.
Robust reliable sampled-data control for switched systems with application to flight control
NASA Astrophysics Data System (ADS)
Sakthivel, R.; Joby, Maya; Shi, P.; Mathiyalagan, K.
2016-11-01
This paper addresses the robust reliable stabilisation problem for a class of uncertain switched systems with random delays and norm bounded uncertainties. The main aim of this paper is to obtain the reliable robust sampled-data control design which involves random time delay with an appropriate gain control matrix for achieving the robust exponential stabilisation for uncertain switched system against actuator failures. In particular, the involved delays are assumed to be randomly time-varying which obeys certain mutually uncorrelated Bernoulli distributed white noise sequences. By constructing an appropriate Lyapunov-Krasovskii functional (LKF) and employing an average-dwell time approach, a new set of criteria is derived for ensuring the robust exponential stability of the closed-loop switched system. More precisely, the Schur complement and Jensen's integral inequality are used in derivation of stabilisation criteria. By considering the relationship among the random time-varying delay and its lower and upper bounds, a new set of sufficient condition is established for the existence of reliable robust sampled-data control in terms of solution to linear matrix inequalities (LMIs). Finally, an illustrative example based on the F-18 aircraft model is provided to show the effectiveness of the proposed design procedures.
Flexible automated approach for quantitative liquid handling of complex biological samples.
Palandra, Joe; Weller, David; Hudson, Gary; Li, Jeff; Osgood, Sarah; Hudson, Emily; Zhong, Min; Buchholz, Lisa; Cohen, Lucinda H
2007-11-01
A fully automated protein precipitation technique for biological sample preparation has been developed for the quantitation of drugs in various biological matrixes. All liquid handling during sample preparation was automated using a Hamilton MicroLab Star Robotic workstation, which included the preparation of standards and controls from a Watson laboratory information management system generated work list, shaking of 96-well plates, and vacuum application. Processing time is less than 30 s per sample or approximately 45 min per 96-well plate, which is then immediately ready for injection onto an LC-MS/MS system. An overview of the process workflow is discussed, including the software development. Validation data are also provided, including specific liquid class data as well as comparative data of automated vs manual preparation using both quality controls and actual sample data. The efficiencies gained from this automated approach are described.
NASA Astrophysics Data System (ADS)
Breton, D. J.; Koffman, B. G.; Kreutz, K. J.; Hamilton, G. S.
2010-12-01
Paleoclimate data are often extracted from ice cores by careful geochemical analysis of meltwater samples. The analysis of the microparticles found in ice cores can also yield unique clues about atmospheric dust loading and transport, dust provenance and past environmental conditions. Determination of microparticle concentration, size distribution and chemical makeup as a function of depth is especially difficult because the particle size measurement either consumes or contaminates the meltwater, preventing further geochemical analysis. Here we describe a microcontroller-based ice core melting system which allows the collection of separate microparticle and chemistry samples from the same depth intervals in the ice core, while logging and accurately depth-tagging real-time electrical conductivity and particle size distribution data. This system was designed specifically to support microparticle analysis of the WAIS Divide WDC06A deep ice core, but many of the subsystems are applicable to more general ice core melting operations. Major system components include: a rotary encoder to measure ice core melt displacement with 0.1 millimeter accuracy, a meltwater tracking system to assign core depths to conductivity, particle and sample vial data, an optical debubbler level control system to protect the Abakus laser particle counter from damage due to air bubbles, a Rabbit 3700 microcontroller which communicates with a host PC, collects encoder and optical sensor data and autonomously operates Gilson peristaltic pumps and fraction collectors to provide automatic sample handling, melt monitor control software operating on a standard PC allowing the user to control and view the status of the system, data logging software operating on the same PC to collect data from the melting, electrical conductivity and microparticle measurement systems. Because microparticle samples can easily be contaminated, we use optical air bubble sensors and high resolution ice core density profiles to guide the melting process. The combination of these data allow us to analyze melt head performance, minimize outer-to-inner fraction contamination and avoid melt head flooding. The WAIS Melt Monitor system allows the collection of real-time, sub-annual microparticle and electrical conductivity data while producing and storing enough sample for traditional Coulter-Counter particle measurements as well long term acid leaching of bioactive metals (e.g., Fe, Co, Cd, Cu, Zn) prior to chemical analysis.
An overview of the thematic mapper geometric correction system
NASA Technical Reports Server (NTRS)
Beyer, E. P.
1983-01-01
Geometric accuracy specifications for LANDSAT 4 are reviewed and the processing concepts which form the basis of NASA's thematic mapper geometric correction system are summarized for both the flight and ground segments. The flight segment includes the thematic mapper instrument, attitude measurement devices, attitude control, and ephemeris processing. For geometric correction the ground segment uses mirror scan correction data, payload correction data, and control point information to determine where TM detector samples fall on output map projection systems. Then the raw imagery is reformatted and resampled to produce image samples on a selected output projection grid system.
Rakkiyappan, R; Sakthivel, N; Cao, Jinde
2015-06-01
This study examines the exponential synchronization of complex dynamical networks with control packet loss and additive time-varying delays. Additionally, sampled-data controller with time-varying sampling period is considered and is assumed to switch between m different values in a random way with given probability. Then, a novel Lyapunov-Krasovskii functional (LKF) with triple integral terms is constructed and by using Jensen's inequality and reciprocally convex approach, sufficient conditions under which the dynamical network is exponentially mean-square stable are derived. When applying Jensen's inequality to partition double integral terms in the derivation of linear matrix inequality (LMI) conditions, a new kind of linear combination of positive functions weighted by the inverses of squared convex parameters appears. In order to handle such a combination, an effective method is introduced by extending the lower bound lemma. To design the sampled-data controller, the synchronization error system is represented as a switched system. Based on the derived LMI conditions and average dwell-time method, sufficient conditions for the synchronization of switched error system are derived in terms of LMIs. Finally, numerical example is employed to show the effectiveness of the proposed methods. Copyright © 2015 Elsevier Ltd. All rights reserved.
NASA Technical Reports Server (NTRS)
Cramer, Christopher J.; Wright, James D.; Simmons, Scott A.; Bobbitt, Lynn E.; DeMoss, Joshua A.
2015-01-01
The paper will present a brief background of the previous data acquisition system at the National Transonic Facility (NTF) and the reasoning and goals behind the upgrade to the current Test SLATE (Test Software Laboratory and Automated Testing Environments) data acquisition system. The components, performance characteristics, and layout of the Test SLATE system within the NTF control room will be discussed. The development, testing, and integration of Test SLATE within NTF operations will be detailed. The operational capabilities of the system will be outlined including: test setup, instrumentation calibration, automatic test sequencer setup, data recording, communication between data and facility control systems, real time display monitoring, and data reduction. The current operational status of the Test SLATE system and its performance during recent NTF testing will be highlighted including high-speed, frame-by-frame data acquisition with conditional sampling post-processing applied. The paper concludes with current development work on the system including the capability for real-time conditional sampling during data acquisition and further efficiency enhancements to the wind tunnel testing process.
A FORTRAN program for the analysis of linear continuous and sample-data systems
NASA Technical Reports Server (NTRS)
Edwards, J. W.
1976-01-01
A FORTRAN digital computer program which performs the general analysis of linearized control systems is described. State variable techniques are used to analyze continuous, discrete, and sampled data systems. Analysis options include the calculation of system eigenvalues, transfer functions, root loci, root contours, frequency responses, power spectra, and transient responses for open- and closed-loop systems. A flexible data input format allows the user to define systems in a variety of representations. Data may be entered by inputing explicit data matrices or matrices constructed in user written subroutines, by specifying transfer function block diagrams, or by using a combination of these methods.
Design of Control Software for a High-Speed Coherent Doppler Lidar System for CO2 Measurement
NASA Technical Reports Server (NTRS)
Vanvalkenburg, Randal L.; Beyon, Jeffrey Y.; Koch, Grady J.; Yu, Jirong; Singh, Upendra N.; Kavaya, Michael J.
2010-01-01
The design of the software for a 2-micron coherent high-speed Doppler lidar system for CO2 measurement at NASA Langley Research Center is discussed in this paper. The specific strategy and design topology to meet the requirements of the system are reviewed. In order to attain the high-speed digitization of the different types of signals to be sampled on multiple channels, a carefully planned design of the control software is imperative. Samples of digitized data from each channel and their roles in data analysis post processing are also presented. Several challenges of extremely-fast, high volume data acquisition are discussed. The software must check the validity of each lidar return as well as other monitoring channel data in real-time. For such high-speed data acquisition systems, the software is a key component that enables the entire scope of CO2 measurement studies using commercially available system components.
Conic Sector Analysis of Hybrid Control Systems. Ph.D. Thesis
NASA Technical Reports Server (NTRS)
Thompson, P. M.
1982-01-01
A hybrid control system contains an analog plant and a hybrid (or sampled-data) compensator. In this thesis a new conic sector is determined which is constructive and can be used to: (1) determine closed loop stability, (2) analyze robustness with respect to modelling uncertainties, (3) analyze steady state response to commands, and (4) select the sample rate. The use of conic sectors allows the designer to treat hybrid control systems as though they were analog control systems. The center of the conic sector can be used as a rigorous linear time invariant approximation of the hybrid control system, and the radius places a bound on the errors of this approximation. The hybrid feedback system can be multivariable, and the sampler is assumed to be synchronous. Algorithms to compute the conic sector are presented. Several examples demonstrate how the conic sector analysis techniques are applied. Extensions to single loop multirate hybrid feedback systems are presented. Further extensions are proposed for multiloop multirate hybrid feedback system and for single rate systems with asynchronous sampling.
NASA Technical Reports Server (NTRS)
Gisser, D. G.; Frederick, D. K.; Lashmet, P. K.; Sandor, G. N.; Shen, C. N.; Yerazunis, S. Y.
1975-01-01
Problems related to an unmanned exploration of the planet Mars by means of an autonomous roving planetary vehicle are investigated. These problems include: design, construction and evaluation of the vehicle itself and its control and operating systems. More specifically, vehicle configuration, dynamics, control, propulsion, hazard detection systems, terrain sensing and modelling, obstacle detection concepts, path selection, decision-making systems, and chemical analyses of samples are studied. Emphasis is placed on development of a vehicle capable of gathering specimens and data for an Augmented Viking Mission or to provide the basis for a Sample Return Mission.
NASA Astrophysics Data System (ADS)
Glatter, Otto; Fuchs, Heribert; Jorde, Christian; Eigner, Wolf-Dieter
1987-03-01
The microprocessor of an 8-bit PC system is used as a central control unit for the acquisition and evaluation of data from quasi-elastic light scattering experiments. Data are sampled with a width of 8 bits under control of the CPU. This limits the minimum sample time to 20 μs. Shorter sample times would need a direct memory access channel. The 8-bit CPU can address a 64-kbyte RAM without additional paging. Up to 49 000 sample points can be measured without interruption. After storage, a correlation function or a power spectrum can be calculated from such a primary data set. Furthermore access is provided to the primary data for stability control, statistical tests, and for comparison of different evaluation methods for the same experiment. A detailed analysis of the signal (histogram) and of the effect of overflows is possible and shows that the number of pulses but not the number of overflows determines the error in the result. The correlation function can be computed with reasonable accuracy from data with a mean pulse rate greater than one, the power spectrum needs a three times higher pulse rate for convergence. The statistical accuracy of the results from 49 000 sample points is of the order of a few percent. Additional averages are necessary to improve their quality. The hardware extensions for the PC system are inexpensive. The main disadvantage of the present system is the high minimum sampling time of 20 μs and the fact that the correlogram or the power spectrum cannot be computed on-line as it can be done with hardware correlators or spectrum analyzers. These shortcomings and the storage size restrictions can be removed with a faster 16/32-bit CPU.
Automated mass spectrometer analysis system
NASA Technical Reports Server (NTRS)
Giffin, Charles E. (Inventor); Kuppermann, Aron (Inventor); Dreyer, William J. (Inventor); Boettger, Heinz G. (Inventor)
1982-01-01
An automated mass spectrometer analysis system is disclosed, in which samples are automatically processed in a sample processor and converted into volatilizable samples, or their characteristic volatilizable derivatives. Each volatilizable sample is sequentially volatilized and analyzed in a double focusing mass spectrometer, whose output is in the form of separate ion beams all of which are simultaneously focused in a focal plane. Each ion beam is indicative of a different sample component or different fragments of one or more sample components and the beam intensity is related to the relative abundance of the sample component. The system includes an electro-optical ion detector which automatically and simultaneously converts the ion beams, first into electron beams which in turn produce a related image which is transferred to the target of a vilicon unit. The latter converts the images into electrical signals which are supplied to a data processor, whose output is a list of the components of the analyzed sample and their abundances. The system is under the control of a master control unit, which in addition to monitoring and controlling various power sources, controls the automatic operation of the system under expected and some unexpected conditions and further protects various critical parts of the system from damage due to particularly abnormal conditions.
Automated mass spectrometer analysis system
NASA Technical Reports Server (NTRS)
Boettger, Heinz G. (Inventor); Giffin, Charles E. (Inventor); Dreyer, William J. (Inventor); Kuppermann, Aron (Inventor)
1978-01-01
An automated mass spectrometer analysis system is disclosed, in which samples are automatically processed in a sample processor and converted into volatilizable samples, or their characteristic volatilizable derivatives. Each volatizable sample is sequentially volatilized and analyzed in a double focusing mass spectrometer, whose output is in the form of separate ion beams all of which are simultaneously focused in a focal plane. Each ion beam is indicative of a different sample component or different fragments of one or more sample components and the beam intensity is related to the relative abundance of the sample component. The system includes an electro-optical ion detector which automatically and simultaneously converts the ion beams, first into electron beams which in turn produce a related image which is transferred to the target of a vidicon unit. The latter converts the images into electrical signals which are supplied to a data processor, whose output is a list of the components of the analyzed sample and their abundances. The system is under the control of a master control unit, which in addition to monitoring and controlling various power sources, controls the automatic operation of the system under expected and some unexpected conditions and further protects various critical parts of the system from damage due to particularly abnormal conditions.
Tsai, Jason Sheng-Hong; Du, Yan-Yi; Huang, Pei-Hsiang; Guo, Shu-Mei; Shieh, Leang-San; Chen, Yuhua
2011-07-01
In this paper, a digital redesign methodology of the iterative learning-based decentralized adaptive tracker is proposed to improve the dynamic performance of sampled-data linear large-scale control systems consisting of N interconnected multi-input multi-output subsystems, so that the system output will follow any trajectory which may not be presented by the analytic reference model initially. To overcome the interference of each sub-system and simplify the controller design, the proposed model reference decentralized adaptive control scheme constructs a decoupled well-designed reference model first. Then, according to the well-designed model, this paper develops a digital decentralized adaptive tracker based on the optimal analog control and prediction-based digital redesign technique for the sampled-data large-scale coupling system. In order to enhance the tracking performance of the digital tracker at specified sampling instants, we apply the iterative learning control (ILC) to train the control input via continual learning. As a result, the proposed iterative learning-based decentralized adaptive tracker not only has robust closed-loop decoupled property but also possesses good tracking performance at both transient and steady state. Besides, evolutionary programming is applied to search for a good learning gain to speed up the learning process of ILC. Copyright © 2011 ISA. Published by Elsevier Ltd. All rights reserved.
Sampled-data controller implementation
NASA Astrophysics Data System (ADS)
Wang, Yu; Leduc, Ryan J.
2012-09-01
The setting of this article is the implementation of timed discrete-event systems (TDES) as sampled-data (SD) controllers. An SD controller is driven by a periodic clock and sees the system as a series of inputs and outputs. On each clock edge (tick event), it samples its inputs, changes states and updates its outputs. In this article, we establish a formal representation of an SD controller as a Moore synchronous finite state machine (FSM). We describe how to translate a TDES supervisor to an FSM, as well as necessary properties to be able to do so. We discuss how to construct a single centralised controller as well as a set of modular controllers, and show that they will produce equivalent output. We briefly discuss how the recently introduced SD controllability definition relates to our translation method. SD controllability is an extension of TDES controllability which captures several new properties that are useful in dealing with concurrency issues, as well as make it easier to translate a TDES supervisor into an SD controller. We next discuss the application of SD controllability to a small flexible manufacturing system (FMS) from the literature. The example demonstrates the successful application of the new SD properties. We describe the design of the system in detail to illustrate the new conditions and to provide designers with guidance on how to apply the properties. We also present some FSM translation issues encountered, as well as the FSM version of the system's supervisors.
An automated atmospheric sampling system operating on 747 airliners
NASA Technical Reports Server (NTRS)
Perkins, P.; Gustafsson, U. R. C.
1975-01-01
An air sampling system that automatically measures the temporal and spatial distribution of selected particulate and gaseous constituents of the atmosphere has been installed on a number of commercial airliners and is collecting data on commercial air routes covering the world. Measurements of constituents related to aircraft engine emissions and other pollutants are made in the upper troposphere and lower stratosphere (6 to 12 km) in support of the Global Air Sampling Program (GASP). Aircraft operated by different airlines sample air at latitudes from the Arctic to Australia. This system includes specialized instrumentation for measuring carbon monoxide, ozone, water vapor, and particulates, a special air inlet probe for sampling outside air, a computerized automatic control, and a data acquisition system. Air constituents and related flight data are tape recorded in flight for later computer processing on the ground.
Baedecker, P.A.; Grossman, J.N.
1995-01-01
A PC based system has been developed for the analysis of gamma-ray spectra and for the complete reduction of data from INAA experiments, including software to average the results from mulitple lines and multiple countings and to produce a final report of analysis. Graphics algorithms may be called for the analysis of complex spectral features, to compare the data from alternate photopeaks and to evaluate detector performance during a given counting cycle. A database of results for control samples can be used to prepare quality control charts to evaluate long term precision and to search for systemic variations in data on reference samples as a function of time. The entire software library can be accessed through a user-friendly menu interface with internal help.
Microcomputer data acquisition and control.
East, T D
1986-01-01
In medicine and biology there are many tasks that involve routine well defined procedures. These tasks are ideal candidates for computerized data acquisition and control. As the performance of microcomputers rapidly increases and cost continues to go down the temptation to automate the laboratory becomes great. To the novice computer user the choices of hardware and software are overwhelming and sadly most of the computer sales persons are not at all familiar with real-time applications. If you want to bill your patients you have hundreds of packaged systems to choose from; however, if you want to do real-time data acquisition the choices are very limited and confusing. The purpose of this chapter is to provide the novice computer user with the basics needed to set up a real-time data acquisition system with the common microcomputers. This chapter will cover the following issues necessary to establish a real time data acquisition and control system: Analysis of the research problem: Definition of the problem; Description of data and sampling requirements; Cost/benefit analysis. Choice of Microcomputer hardware and software: Choice of microprocessor and bus structure; Choice of operating system; Choice of layered software. Digital Data Acquisition: Parallel Data Transmission; Serial Data Transmission; Hardware and software available. Analog Data Acquisition: Description of amplitude and frequency characteristics of the input signals; Sampling theorem; Specification of the analog to digital converter; Hardware and software available; Interface to the microcomputer. Microcomputer Control: Analog output; Digital output; Closed-Loop Control. Microcomputer data acquisition and control in the 21st Century--What is in the future? High speed digital medical equipment networks; Medical decision making and artificial intelligence.
Factors of quality of financial report of local government in Indonesia
NASA Astrophysics Data System (ADS)
Muda, Iskandar; Haris Harahap, Abdul; Erlina; Ginting, Syafruddin; Maksum, Azhar; Abubakar, Erwin
2018-03-01
The purpose of this research is to find out whether the Accounting Information System and Internal Control in Local Revenue Office to the affect the Quality of Financial Report of Local Government. The sampling was conducted by using simple random sampling method in which the sample was determined without considering strata. The data research was conducted by distributing the questionnaires. The results showed that the degree of Accounting Information System and Internal Control simultaneously affect the Quality of Financial Report of Local Government. However, partially, Partially, accounting information system influence to the quality of financial report of local government and the internal control does not affect the quality of financial report.
Consensus for second-order multi-agent systems with position sampled data
NASA Astrophysics Data System (ADS)
Wang, Rusheng; Gao, Lixin; Chen, Wenhai; Dai, Dameng
2016-10-01
In this paper, the consensus problem with position sampled data for second-order multi-agent systems is investigated. The interaction topology among the agents is depicted by a directed graph. The full-order and reduced-order observers with position sampled data are proposed, by which two kinds of sampled data-based consensus protocols are constructed. With the provided sampled protocols, the consensus convergence analysis of a continuous-time multi-agent system is equivalently transformed into that of a discrete-time system. Then, by using matrix theory and a sampled control analysis method, some sufficient and necessary consensus conditions based on the coupling parameters, spectrum of the Laplacian matrix and sampling period are obtained. While the sampling period tends to zero, our established necessary and sufficient conditions are degenerated to the continuous-time protocol case, which are consistent with the existing result for the continuous-time case. Finally, the effectiveness of our established results is illustrated by a simple simulation example. Project supported by the Natural Science Foundation of Zhejiang Province, China (Grant No. LY13F030005) and the National Natural Science Foundation of China (Grant No. 61501331).
The Galileo scan platform pointing control system - A modern control theoretic viewpoint
NASA Technical Reports Server (NTRS)
Sevaston, G. E.; Macala, G. A.; Man, G. K.
1985-01-01
The current Galileo scan platform pointing control system (SPPCS) is described, and ways in which modern control concepts could serve to enhance it are considered. Of particular interest are: the multi-variable design model and overall control system architecture, command input filtering, feedback compensator and command input design, stability robustness constraint for both continuous time control systems and for sampled data control systems, and digital implementation of the control system. The proposed approach leads to the design of a system that is similar to current Galileo SPPCS configuration, but promises to be more systematic.
NASA Astrophysics Data System (ADS)
Kim, Jung Hoon; Hagiwara, Tomomichi
2017-11-01
This paper is concerned with linear time-invariant (LTI) sampled-data systems (by which we mean sampled-data systems with LTI generalised plants and LTI controllers) and studies their H2 norms from the viewpoint of impulse responses and generalised H2 norms from the viewpoint of the induced norms from L2 to L∞. A new definition of the H2 norm of LTI sampled-data systems is first introduced through a sort of intermediate standpoint of those for the existing two definitions. We then establish unified treatment of the three definitions of the H2 norm through a matrix function G(τ) defined on the sampling interval [0, h). This paper next considers the generalised H2 norms, in which two types of the L∞ norm of the output are considered as the temporal supremum magnitude under the spatial 2-norm and ∞-norm of a vector-valued function. We further give unified treatment of the generalised H2 norms through another matrix function F(θ) which is also defined on [0, h). Through a close connection between G(τ) and F(θ), some theoretical relationships between the H2 and generalised H2 norms are provided. Furthermore, appropriate extensions associated with the treatment of G(τ) and F(θ) to the closed interval [0, h] are discussed to facilitate numerical computations and comparisons of the H2 and generalised H2 norms. Through theoretical and numerical studies, it is shown that the two generalised H2 norms coincide with neither of the three H2 norms of LTI sampled-data systems even though all the five definitions coincide with each other when single-output continuous-time LTI systems are considered as a special class of LTI sampled-data systems. To summarise, this paper clarifies that the five control performance measures are mutually related with each other but they are also intrinsically different from each other.
AN AUTOMATED MONITORING SYSTEM FOR FISH PHYSIOLOGY AND TOXICOLOGY
This report describes a data acquisition and control (DAC) system that was constructed to manage selected physiological measurements and sample control for aquatic physiology and toxicology. Automated DAC was accomplished with a microcomputer running menu-driven software develope...
The Development and Support of the NATO Project OPAQUE U.S.A.F. System Control Programs.
1978-12-30
OPAQUE data bank format. D r Di,3t SECURITY CLASSIFICATION OF THIS PAGE(Whet, Date Entoed) TABLE OF CONTENTS A ckn ol edgem en-s--------- ------- - - vi 1...Figure 11.0. ERIKFILE Contents Summary Sample (ERIK)------------------------- 64. Figure II.P. ERIKFILE Hourly Data Dump Sample (ERIK...sequentiallv and executed by a control card call to the procedure file. [lhe contents of the monthly stripped data files and their associated directories
Wu, Zheng-Guang; Shi, Peng; Su, Hongye; Chu, Jian
2012-09-01
This paper investigates the problem of master-slave synchronization for neural networks with discrete and distributed delays under variable sampling with a known upper bound on the sampling intervals. An improved method is proposed, which captures the characteristic of sampled-data systems. Some delay-dependent criteria are derived to ensure the exponential stability of the error systems, and thus the master systems synchronize with the slave systems. The desired sampled-data controller can be achieved by solving a set of linear matrix inequalitys, which depend upon the maximum sampling interval and the decay rate. The obtained conditions not only have less conservatism but also have less decision variables than existing results. Simulation results are given to show the effectiveness and benefits of the proposed methods.
NASA Technical Reports Server (NTRS)
Wallace, J. W.; Lovelady, R. W.; Ferguson, R. L.
1981-01-01
A prototype water quality monitoring system is described which offers almost continuous in situ monitoring. The two-man portable system features: (1) a microprocessor controlled central processing unit which allows preprogrammed sampling schedules and reprogramming in situ; (2) a subsurface unit for multiple depth capability and security from vandalism; (3) an acoustic data link for communications between the subsurface unit and the surface control unit; (4) eight water quality parameter sensors; (5) a nonvolatile magnetic bubble memory which prevents data loss in the event of power interruption; (6) a rechargeable power supply sufficient for 2 weeks of unattended operation; (7) a water sampler which can collect samples for laboratory analysis; (8) data output in direct engineering units on printed tape or through a computer compatible link; (9) internal electronic calibration eliminating external sensor adjustment; and (10) acoustic location and recovery systems. Data obtained in Saginaw Bay, Lake Huron are tabulated.
Fiber-optic temperature probe system for inner body
NASA Astrophysics Data System (ADS)
Liu, Bo; Deng, Xing-Zhong; Cao, Wei; Cheng, Xianping; Xie, Tuqiang; Zhong, Zugen
1991-08-01
The authors have designed a fiber-optic temperature probe system that can quickly insert its probe into bodies to measure temperature. Its thermometer unit has the function of program- controlled zeroing. The single-chip microcomputer is used to control the whole system and process data. The sample system has been tested in a coal furnace.
Imaging system design and image interpolation based on CMOS image sensor
NASA Astrophysics Data System (ADS)
Li, Yu-feng; Liang, Fei; Guo, Rui
2009-11-01
An image acquisition system is introduced, which consists of a color CMOS image sensor (OV9620), SRAM (CY62148), CPLD (EPM7128AE) and DSP (TMS320VC5509A). The CPLD implements the logic and timing control to the system. SRAM stores the image data, and DSP controls the image acquisition system through the SCCB (Omni Vision Serial Camera Control Bus). The timing sequence of the CMOS image sensor OV9620 is analyzed. The imaging part and the high speed image data memory unit are designed. The hardware and software design of the image acquisition and processing system is given. CMOS digital cameras use color filter arrays to sample different spectral components, such as red, green, and blue. At the location of each pixel only one color sample is taken, and the other colors must be interpolated from neighboring samples. We use the edge-oriented adaptive interpolation algorithm for the edge pixels and bilinear interpolation algorithm for the non-edge pixels to improve the visual quality of the interpolated images. This method can get high processing speed, decrease the computational complexity, and effectively preserve the image edges.
Engineering Design of ITER Prototype Fast Plant System Controller
NASA Astrophysics Data System (ADS)
Goncalves, B.; Sousa, J.; Carvalho, B.; Rodrigues, A. P.; Correia, M.; Batista, A.; Vega, J.; Ruiz, M.; Lopez, J. M.; Rojo, R. Castro; Wallander, A.; Utzel, N.; Neto, A.; Alves, D.; Valcarcel, D.
2011-08-01
The ITER control, data access and communication (CODAC) design team identified the need for two types of plant systems. A slow control plant system is based on industrial automation technology with maximum sampling rates below 100 Hz, and a fast control plant system is based on embedded technology with higher sampling rates and more stringent real-time requirements than that required for slow controllers. The latter is applicable to diagnostics and plant systems in closed-control loops whose cycle times are below 1 ms. Fast controllers will be dedicated industrial controllers with the ability to supervise other fast and/or slow controllers, interface to actuators and sensors and, if necessary, high performance networks. Two prototypes of a fast plant system controller specialized for data acquisition and constrained by ITER technological choices are being built using two different form factors. This prototyping activity contributes to the Plant Control Design Handbook effort of standardization, specifically regarding fast controller characteristics. Envisaging a general purpose fast controller design, diagnostic use cases with specific requirements were analyzed and will be presented along with the interface with CODAC and sensors. The requirements and constraints that real-time plasma control imposes on the design were also taken into consideration. Functional specifications and technology neutral architecture, together with its implications on the engineering design, were considered. The detailed engineering design compliant with ITER standards was performed and will be discussed in detail. Emphasis will be given to the integration of the controller in the standard CODAC environment. Requirements for the EPICS IOC providing the interface to the outside world, the prototype decisions on form factor, real-time operating system, and high-performance networks will also be discussed, as well as the requirements for data streaming to CODAC for visualization and archiving.
A controlled rate freeze/thaw system for cryopreservation of biological materials
NASA Technical Reports Server (NTRS)
Anselmo, V. J.; Harrison, R. G.
1979-01-01
A system which allows programmable temperature-time control for a 5 cc sample volume of an arbitrary biological material was constructed. Steady state and dynamic temperature control was obtained by supplying heat to the sample volume through resistive elements constructed as an integral part of the sample container. For cooling purposes, this container was totally immersed into a cold heat sink. Sample volume thermodynamic property data were obtained by measurements of heater power and heat flux through the container walls. Using a mixture of dry ice and alcohol at -79 C, sample volume was controlled from +40 C to -60 C at rates from steady state to + or - 65 C/min. Steady state temperature precision was better than 0.2 C while the dynamic capability depends on the temperature rate of change as well as the thermal mass of the sample and the container.
Analyzing Feedback Control Systems
NASA Technical Reports Server (NTRS)
Bauer, Frank H.; Downing, John P.
1987-01-01
Interactive controls analysis (INCA) program developed to provide user-friendly environment for design and analysis of linear control systems, primarily feedback control. Designed for use with both small- and large-order systems. Using interactive-graphics capability, INCA user quickly plots root locus, frequency response, or time response of either continuous-time system or sampled-data system. Configuration and parameters easily changed, allowing user to design compensation networks and perform sensitivity analyses in very convenient manner. Written in Pascal and FORTRAN.
DAQ application of PC oscilloscope for chaos fiber-optic fence system based on LabVIEW
NASA Astrophysics Data System (ADS)
Lu, Manman; Fang, Nian; Wang, Lutang; Huang, Zhaoming; Sun, Xiaofei
2011-12-01
In order to obtain simultaneously high sample rate and large buffer in data acquisition (DAQ) for a chaos fiber-optic fence system, we developed a double-channel high-speed DAQ application of a digital oscilloscope of PicoScope 5203 based on LabVIEW. We accomplished it by creating call library function (CLF) nodes to call the DAQ functions in the two dynamic link libraries (DLLs) of PS5000.dll and PS5000wrap.dll provided by Pico Technology Company. The maximum real-time sample rate of the DAQ application can reach 1GS/s. We can control the resolutions of the application at the sample time and data amplitudes by changing their units in the block diagram, and also control the start and end times of the sampling operations. The experimental results show that the application has enough high sample rate and large buffer to meet the demanding DAQ requirements of the chaos fiber-optic fence system.
Two-dimensional simple proportional feedback control of a chaotic reaction system
NASA Astrophysics Data System (ADS)
Mukherjee, Ankur; Searson, Dominic P.; Willis, Mark J.; Scott, Stephen K.
2008-04-01
The simple proportional feedback (SPF) control algorithm may, in principle, be used to attain periodic oscillations in dynamic systems exhibiting low-dimensional chaos. However, if implemented within a discrete control framework with sampling frequency limitations, controller performance may deteriorate. This phenomenon is illustrated using simulations of a chaotic autocatalytic reaction system. A two-dimensional (2D) SPF controller that explicitly takes into account some of the problems caused by limited sampling rates is then derived by introducing suitable modifications to the original SPF method. Using simulations, the performance of the 2D-SPF controller is compared to that of a conventional SPF control law when implemented as a sampled data controller. Two versions of the 2D-SPF controller are described: linear (L2D-SPF) and quadratic (Q2D-SPF). The performance of both the L2D-SPF and Q2D-SPF controllers is shown to be superior to the SPF when controller sampling frequencies are decreased. Furthermore, it is demonstrated that the Q2D-SPF controller provides better fixed point stabilization compared to both the L2D-SPF and the conventional SPF when concentration measurements are corrupted by noise.
Randomly Sampled-Data Control Systems. Ph.D. Thesis
NASA Technical Reports Server (NTRS)
Han, Kuoruey
1990-01-01
The purpose is to solve the Linear Quadratic Regulator (LQR) problem with random time sampling. Such a sampling scheme may arise from imperfect instrumentation as in the case of sampling jitter. It can also model the stochastic information exchange among decentralized controllers to name just a few. A practical suboptimal controller is proposed with the nice property of mean square stability. The proposed controller is suboptimal in the sense that the control structure is limited to be linear. Because of i. i. d. assumption, this does not seem unreasonable. Once the control structure is fixed, the stochastic discrete optimal control problem is transformed into an equivalent deterministic optimal control problem with dynamics described by the matrix difference equation. The N-horizon control problem is solved using the Lagrange's multiplier method. The infinite horizon control problem is formulated as a classical minimization problem. Assuming existence of solution to the minimization problem, the total system is shown to be mean square stable under certain observability conditions. Computer simulations are performed to illustrate these conditions.
Sampling and Control Circuit Board for an Inertial Measurement Unit
NASA Technical Reports Server (NTRS)
Chelmins, David T (Inventor); Sands, Obed (Inventor); Powis, Richard T., Jr. (Inventor)
2016-01-01
A circuit board that serves as a control and sampling interface to an inertial measurement unit ("IMU") is provided. The circuit board is also configured to interface with a local oscillator and an external trigger pulse. The circuit board is further configured to receive the external trigger pulse from an external source that time aligns the local oscillator and initiates sampling of the inertial measurement device for data at precise time intervals based on pulses from the local oscillator. The sampled data may be synchronized by the circuit board with other sensors of a navigation system via the trigger pulse.
Serial data correlator/code translator
NASA Technical Reports Server (NTRS)
Morgan, L. E. (Inventor)
1982-01-01
A system for analyzing asynchronous signals containing bits of information for ensuring the validity of said signals, by sampling each bit of information a plurality of times, and feeding the sampled pieces of bits of information into a sequence controlled is described. The sequence controller has a plurality of maps or programs through which the sampled pieces of bits are stepped so as to identify the particular bit of information and determine the validity and phase of the bit. The step in which the sequence controller is clocked is controlled by a storage register. A data decoder decodes the information fed out of the storage register and feeds such information to shift registers for storage.
Chaos: Understanding and Controlling Laser Instability
NASA Technical Reports Server (NTRS)
Blass, William E.
1997-01-01
In order to characterize the behavior of tunable diode lasers (TDL), the first step in the project involved the redesign of the TDL system here at the University of Tennessee Molecular Systems Laboratory (UTMSL). Having made these changes it was next necessary to optimize the new optical system. This involved the fine adjustments to the optical components, particularly in the monochromator, to minimize the aberrations of coma and astigmatism and to assure that the energy from the beam is focused properly on the detector element. The next step involved the taking of preliminary data. We were then ready for the analysis of the preliminary data. This required the development of computer programs that use mathematical techniques to look for signatures of chaos. Commercial programs were also employed. We discovered some indication of high dimensional chaos, but were hampered by the low sample rate of 200 KSPS (kilosamples/sec) and even more by our sample size of 1024 (1K) data points. These limitations were expected and we added a high speed data acquisition board. We incorporated into the system a computer with a 40 MSPS (million samples/sec) data acquisition board. This board can also capture 64K of data points so that were then able to perform the more accurate tests for chaos. The results were dramatic and compelling, we had demonstrated that the lead salt diode laser had a chaotic frequency output. Having identified the chaotic character in our TDL data, we proceeded to stage two as outlined in our original proposal. This required the use of an Occasional Proportional Feedback (OPF) controller to facilitate the control and stabilization of the TDL system output. The controller was designed and fabricated at GSFC and debugged in our laboratories. After some trial and error efforts, we achieved chaos control of the frequency emissions of the laser. The two publications appended to this introduction detail the entire project and its results.
Research study on IPS digital controller design
NASA Technical Reports Server (NTRS)
Kuo, B. C.; Folkerts, C.
1976-01-01
The performance is investigated of the simplified continuous-data model of the Instrument Pointing System (IPS). Although the ultimate objective is to study the digital model of the system, knowledge on the performance of the continuous-data model is important in the sense that the characteristics of the digital system should approach those of the continuous-data system as the sampling period approaches zero.
NASA Technical Reports Server (NTRS)
Peters, P. N.; Hester, H. B.; Bertsch, W.; Mayfield, H.; Zatko, D.
1983-01-01
An investigation involving sampling the rapidly changing environment of the Shuttle cargo bay is considered. Four time-integrated samples and one rapid acquisition sample were collected to determine the types and quantities of contaminants present during ascent and descent of the Shuttle. The sampling times for the various bottles were controlled by valves operated by the Data Acquisition and Control System (DACS) of the IECM. Many of the observed species were found to be common solvents used in cleaning surfaces. When the actual volume sampled is taken into account, the relative mass of organics sampled during descent is about 20 percent less than during ascent.
Wang, Tong; Gao, Huijun; Qiu, Jianbin
2016-02-01
This paper investigates the multirate networked industrial process control problem in double-layer architecture. First, the output tracking problem for sampled-data nonlinear plant at device layer with sampling period T(d) is investigated using adaptive neural network (NN) control, and it is shown that the outputs of subsystems at device layer can track the decomposed setpoints. Then, the outputs and inputs of the device layer subsystems are sampled with sampling period T(u) at operation layer to form the index prediction, which is used to predict the overall performance index at lower frequency. Radial basis function NN is utilized as the prediction function due to its approximation ability. Then, considering the dynamics of the overall closed-loop system, nonlinear model predictive control method is proposed to guarantee the system stability and compensate the network-induced delays and packet dropouts. Finally, a continuous stirred tank reactor system is given in the simulation part to demonstrate the effectiveness of the proposed method.
ROMPS critical design review data package
NASA Technical Reports Server (NTRS)
Dobbs, M. E.
1992-01-01
The design elements of the Robot-Operated Material Processing in Space (ROMPS) system are described in outline and graphical form. The following subsystems/topics are addressed: servo system, testbed and simulation results, System V Controller, robot module, furnace module, SCL experiment supervisor and script sample processing control, battery system, watchdog timers, mechanical/thermal considerations, and fault conditions and recovery.
Lincoln, Tricia A.; Horan-Ross, Debra A.; McHale, Michael R.; Lawrence, Gregory B.
2006-01-01
The laboratory for analysis of low-ionic-strength water at the U.S. Geological Survey (USGS) Water Science Center in Troy, N.Y., analyzes samples collected by USGS projects throughout the Northeast. The laboratory's quality-assurance program is based on internal and interlaboratory quality-assurance samples and quality-control procedures that were developed to ensure proper sample collection, processing, and analysis. The quality-assurance and quality-control data were stored in the laboratory's LabMaster data-management system, which provides efficient review, compilation, and plotting of data. This report presents and discusses results of quality-assurance and quality-control samples analyzed from July 1999 through June 2001. Results for the quality-control samples for 18 analytical procedures were evaluated for bias and precision. Control charts indicate that data for eight of the analytical procedures were occasionally biased for either high-concentration or low-concentration samples but were within control limits; these procedures were: acid-neutralizing capacity, total monomeric aluminum, total aluminum, calcium, chloride and nitrate (ion chromatography and colormetric method) and sulfate. The total aluminum and dissolved organic carbon procedures were biased throughout the analysis period for the high-concentration sample, but were within control limits. The calcium and specific conductance procedures were biased throughout the analysis period for the low-concentration sample, but were within control limits. The magnesium procedure was biased for the high-concentration and low concentration samples, but was within control limits. Results from the filter-blank and analytical-blank analyses indicate that the procedures for 14 of 15 analytes were within control limits, although the concentrations for blanks were occasionally outside the control limits. The data-quality objective was not met for dissolved organic carbon. Sampling and analysis precision are evaluated herein in terms of the coefficient of variation obtained for triplicate samples in the procedures for 17 of the 18 analytes. At least 90 percent of the samples met data-quality objectives for all analytes except ammonium (81 percent of samples met objectives), chloride (75 percent of samples met objectives), and sodium (86 percent of samples met objectives). Results of the USGS interlaboratory Standard Reference Sample (SRS) Project indicated good data quality over the time period, with most ratings for each sample in the good to excellent range. The P-sample (low-ionic-strength constituents) analysis had one satisfactory rating for the specific conductance procedure in one study. The T-sample (trace constituents) analysis had one satisfactory rating for the aluminum procedure in one study and one unsatisfactory rating for the sodium procedure in another. The remainder of the samples had good or excellent ratings for each study. Results of Environment Canada's National Water Research Institute (NWRI) program indicated that at least 89 percent of the samples met data-quality objectives for 10 of the 14 analytes; the exceptions were ammonium, total aluminum, dissolved organic carbon, and sodium. Results indicate a positive bias for the ammonium procedure in all studies. Data-quality objectives were not met in 50 percent of samples analyzed for total aluminum, 38 percent of samples analyzed for dissolved organic carbon, and 27 percent of samples analyzed for sodium. Results from blind reference-sample analyses indicated that data-quality objectives were met by at least 91 percent of the samples analyzed for calcium, chloride, fluoride, magnesium, pH, potassium, and sulfate. Data-quality objectives were met by 75 percent of the samples analyzed for sodium and 58 percent of the samples analyzed for specific conductance.
NASA Astrophysics Data System (ADS)
Shan, Xuchen; Zhang, Bei; Lan, Guoqiang; Wang, Yiqiao; Liu, Shugang
2015-11-01
Biology and medicine sample measurement takes an important role in the microscopic optical technology. Optical tweezer has the advantage of accurate capture and non-pollution of the sample. The SPR(surface plasmon resonance) sensor has so many advantages include high sensitivity, fast measurement, less consumption of sample and label-free detection of biological sample that the SPR sensing technique has been used for surface topography, analysis of biochemical and immune, drug screening and environmental monitoring. If they combine, they will play an important role in the biological, chemical and other subjects. The system we propose use the multi-axis cage system, by using the methods of reflection and transmiss ion to improve the space utilization. The SPR system and optical tweezer were builtup and combined in one system. The cage of multi-axis system gives full play to its accuracy, simplicity and flexibility. The size of the system is 20 * 15 * 40 cm3 and thus the sample can be replaced to switch between the optical tweezers system and the SPR system in the small space. It means that we get the refractive index of the sample and control the particle in the same system. In order to control the revolving stage, get the picture and achieve the data stored automatically, we write a LabVIEW procedure. Then according to the data from the back focal plane calculate the refractive index of the sample. By changing the slide we can trap the particle as optical tweezer, which makes us measurement and trap the sample at the same time.
Monitoring service for the Gran Telescopio Canarias control system
NASA Astrophysics Data System (ADS)
Huertas, Manuel; Molgo, Jordi; Macías, Rosa; Ramos, Francisco
2016-07-01
The Monitoring Service collects, persists and propagates the Telescope and Instrument telemetry, for the Gran Telescopio CANARIAS (GTC), an optical-infrared 10-meter segmented mirror telescope at the ORM observatory in Canary Islands (Spain). A new version of the Monitoring Service has been developed in order to improve performance, provide high availability, guarantee fault tolerance and scalability to cope with high volume of data. The architecture is based on a distributed in-memory data store with a Product/Consumer pattern design. The producer generates the data samples. The consumers either persists the samples to a database for further analysis or propagates them to the consoles in the control room to monitorize the state of the whole system.
Lincoln, Tricia A.; Horan-Ross, Debra A.; McHale, Michael R.; Lawrence, Gregory B.
2001-01-01
A laboratory for analysis of low-ionic strength water has been developed at the U.S. Geological Survey (USGS) office in Troy, N.Y., to analyze samples collected by USGS projects in the Northeast. The laboratory's quality-assurance program is based on internal and interlaboratory quality-assurance samples and quality-control procedures developed to ensure proper sample collection, processing, and analysis. The quality-assurance/quality-control data are stored in the laboratory's SAS data-management system, which provides efficient review, compilation, and plotting of quality-assurance/quality-control data. This report presents and discusses samples analyzed from July 1993 through June 1995. Quality-control results for 18 analytical procedures were evaluated for bias and precision. Control charts show that data from seven of the analytical procedures were biased throughout the analysis period for either high-concentration or low-concentration samples but were within control limits; these procedures were: acid-neutralizing capacity, dissolved inorganic carbon, dissolved organic carbon (soil expulsions), chloride, magnesium, nitrate (colorimetric method), and pH. Three of the analytical procedures were occasionally biased but were within control limits; they were: calcium (high for high-concentration samples for May 1995), dissolved organic carbon (high for highconcentration samples from January through September 1994), and fluoride (high in samples for April and June 1994). No quality-control sample has been developed for the organic monomeric aluminum procedure. Results from the filter-blank and analytical-blank analyses indicate that all analytical procedures in which blanks were run were within control limits, although values for a few blanks were outside the control limits. Blanks were not analyzed for acid-neutralizing capacity, dissolved inorganic carbon, fluoride, nitrate (colorimetric method), or pH. Sampling and analysis precision are evaluated herein in terms of the coefficient of variation obtained for triplicate samples in 14 of the 18 procedures. Data-quality objectives were met by more than 90 percent of the samples analyzed in all procedures except total monomeric aluminum (85 percent of samples met objectives), total aluminum (70 percent of samples met objectives), and dissolved organic carbon (85 percent of samples met objectives). Triplicate samples were not analyzed for ammonium, fluoride, dissolved inorganic carbon, or nitrate (colorimetric method). Results of the USGS interlaboratory Standard Reference Sample Program indicated high data quality with a median result of 3.6 of a possible 4.0. Environment Canada's LRTAP interlaboratory study results indicated that more than 85 percent of the samples met data-quality objectives in 6 of the 12 analyses; exceptions were calcium, dissolved organic carbon, chloride, pH, potassium, and sodium. Data-quality objectives were not met for calcium samples in one LRTAP study, but 94 percent of samples analyzed were within control limits for the remaining studies. Data-quality objectives were not met by 35 percent of samples analyzed for dissolved organic carbon, but 94 percent of sample values were within 20 percent of the most probable value. Data-quality objectives were not met for 30 percent of samples analyzed for chloride, but 90 percent of sample values were within 20 percent of the most probable value. Measurements of samples with a pH above 6.0 were biased high in 54 percent of the samples, although 85 percent of the samples met data-quality objectives for pH measurements below 6.0. Data-quality objectives for potassium and sodium were not met in one study (only 33 percent of the samples analyzed met the objectives), although 85 percent of the sample values were within control limits for the other studies. Measured sodium values were above the upper control limit in all studies. Results from blind reference-sample analyses indicated that data
THYME: Toolkit for Hybrid Modeling of Electric Power Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nutaro Kalyan Perumalla, James Joseph
2011-01-01
THYME is an object oriented library for building models of wide area control and communications in electric power systems. This software is designed as a module to be used with existing open source simulators for discrete event systems in general and communication systems in particular. THYME consists of a typical model for simulating electro-mechanical transients (e.g., as are used in dynamic stability studies), data handling objects to work with CDF and PTI formatted power flow data, and sample models of discrete sensors and controllers.
NASA Technical Reports Server (NTRS)
Seltzer, S. M.
1976-01-01
The problem discussed is to design a digital controller for a typical satellite. The controlled plant is considered to be a rigid body acting in a plane. The controller is assumed to be a digital computer which, when combined with the proposed control algorithm, can be represented as a sampled-data system. The objective is to present a design strategy and technique for selecting numerical values for the control gains (assuming position, integral, and derivative feedback) and the sample rate. The technique is based on the parameter plane method and requires that the system be amenable to z-transform analysis.
Lincoln, Tricia A.; Horan-Ross, Debra A.; McHale, Michael R.; Lawrence, Gregory B.
2006-01-01
The laboratory for analysis of low-ionic-strength water at the U.S. Geological Survey (USGS) Water Science Center in Troy, N.Y., analyzes samples collected by USGS projects throughout the Northeast. The laboratory's quality-assurance program is based on internal and interlaboratory quality-assurance samples and quality-control procedures that were developed to ensure proper sample collection, processing, and analysis. The quality-assurance/quality-control data for the time period addressed in this report were stored in the laboratory's SAS data-management system, which provides efficient review, compilation, and plotting of data. This report presents and discusses results of quality-assurance and quality- control samples analyzed from July 1997 through June 1999. Results for the quality-control samples for 18 analytical procedures were evaluated for bias and precision. Control charts indicate that data for eight of the analytical procedures were occasionally biased for either high-concentration and (or) low-concentration samples but were within control limits; these procedures were: acid-neutralizing capacity, total monomeric aluminum, total aluminum, ammonium, calcium, chloride, specific conductance, and sulfate. The data from the potassium and sodium analytical procedures are insufficient for evaluation. Results from the filter-blank and analytical-blank analyses indicate that the procedures for 11 of 13 analytes were within control limits, although the concentrations for blanks were occasionally outside the control limits. Blank analysis results for chloride showed that 22 percent of blanks did not meet data-quality objectives and results for dissolved organic carbon showed that 31 percent of the blanks did not meet data-quality objectives. Sampling and analysis precision are evaluated herein in terms of the coefficient of variation obtained for triplicate samples in the procedures for 14 of the 18 analytes. At least 90 percent of the samples met data-quality objectives for all analytes except total aluminum (70 percent of samples met objectives) and potassium (83 percent of samples met objectives). Results of the USGS interlaboratory Standard Reference Sample (SRS) Project indicated good data quality for most constituents over the time period. The P-sample (low-ionic-strength constituents) analysis had good ratings in two of these studies and a satisfactory rating in the third. The results of the T-sample (trace constituents) analysis indicated high data quality with good ratings in all three studies. The N-sample (nutrient constituents) studies had one each of excellent, good, and satisfactory ratings. Results of Environment Canada's National Water Research Institute (NWRI) program indicated that at least 80 percent of the samples met data-quality objectives for 9 of the 13 analytes; the exceptions were dissolved organic carbon, ammonium, chloride, and specific conductance. Data-quality objectives were not met for dissolved organic carbon in two NWRI studies, but all of the samples were within control limits for the last study. Data-quality objectives were not met in 41 percent of samples analyzed for ammonium, 25 percent of samples analyzed for chloride, and 30 percent of samples analyzed for specific conductance. Results from blind reference-sample analyses indicated that data-quality objectives were met by at least 84 percent of the samples analyzed for calcium, chloride, magnesium, pH, and potassium. Data-quality objectives were met by 73 percent of those analyzed for sulfate. The data-quality objective was not met for sodium. The data are insufficient for evaluation of the specific conductance results.
Flight-test experience in digital control of a remotely piloted vehicle.
NASA Technical Reports Server (NTRS)
Edwards, J. W.
1972-01-01
The development of a remotely piloted vehicle system consisting of a remote pilot cockpit and a ground-based digital computer coupled to the aircraft through telemetry data links is described. The feedback control laws are implemented in a FORTRAN program. Flight-test experience involving high feedback gain limits for attitude and attitude rate feedback variables, filtering of sampled data, and system operation during intermittent telemetry data link loss is discussed. Comparisons of closed-loop flight tests with analytical calculations, and pilot comments on system operation are included.
Kuntanapreeda, S; Fullmer, R R
1996-01-01
A training method for a class of neural network controllers is presented which guarantees closed-loop system stability. The controllers are assumed to be nonlinear, feedforward, sampled-data, full-state regulators implemented as single hidden-layer neural networks. The controlled systems must be locally hermitian and observable. Stability of the closed-loop system is demonstrated by determining a Lyapunov function, which can be used to identify a finite stability region about the regulator point.
Zhang, Xian-Ming; Han, Qing-Long
2016-12-01
This paper is concerned with decentralized event-triggered dissipative control for systems with the entries of the system outputs having different physical properties. Depending on these different physical properties, the entries of the system outputs are grouped into multiple nodes. A number of sensors are used to sample the signals from different nodes. A decentralized event-triggering scheme is introduced to select those necessary sampled-data packets to be transmitted so that communication resources can be saved significantly while preserving the prescribed closed-loop performance. First, in order to organize the decentralized data packets transmitted from the sensor nodes, a data packet processor (DPP) is used to generate a new signal to be held by the zero-order-hold once the signal stored by the DPP is updated at some time instant. Second, under the mechanism of the DPP, the resulting closed-loop system is modeled as a linear system with an interval time-varying delay. A sufficient condition is derived such that the closed-loop system is asymptotically stable and strictly (Q 0 ,S 0 ,R 0 ) -dissipative, where Q 0 ,S 0 , and R 0 are real matrices of appropriate dimensions with Q 0 and R 0 symmetric. Third, suitable output-based controllers can be designed based on solutions to a set of a linear matrix inequality. Finally, two examples are given to demonstrate the effectiveness of the proposed method.
NASA Astrophysics Data System (ADS)
Talukder, A.; Panangadan, A. V.; Blumberg, A. F.; Herrington, T.; Georgas, N.
2008-12-01
The New York Harbor Observation and Prediction System (NYHOPS) is a real-time, estuarine and coastal ocean observing and modeling system for the New York Harbor and surrounding waters. Real-time measurements from in-situ mobile and stationary sensors in the NYHOPS networks are assimilated into marine forecasts in order to reduce the discrepancy with ground truth. The forecasts are obtained from the ECOMSED hydrodynamic model, a shallow water derivative of the Princeton Ocean Model. Currently, all sensors in the NYHOPS system are operated in a fixed mode with uniform sampling rates. This technology infusion effort demonstrates the use of Model Predictive Control (MPC) to autonomously adapt the operation of both mobile and stationary sensors in response to changing events that are -automatically detected from the ECOMSED forecasts. The controller focuses sensing resources on those regions that are expected to be impacted by the detected events. The MPC approach involves formulating the problem of calculating the optimal sensor parameters as a constrained multi-objective optimization problem. We have developed an objective function that takes into account the spatiotemporal relationship of the in-situ sensor locations and the locations of events detected by the model. Experiments in simulation were carried out using data collected during a freshwater flooding event. The location of the resulting freshwater plume was calculated from the corresponding model forecasts and was used by the MPC controller to derive control parameters for the sensing assets. The operational parameters that are controlled include the sampling rates of stationary sensors, paths of unmanned underwater vehicles (UUVs), and data transfer routes between sensors and the central modeling computer. The simulation experiments show that MPC-based sensor control reduces the RMS error in the forecast by a factor of 380% as compared to uniform sampling. The paths of multiple UUVs were simultaneously calculated such that measurements from on-board sensors would lead to maximal reduction in the forecast error after data assimilation. The MPC controller also reduces the consumption of system resources such as energy expended in sampling and wireless communication. The MPC-based control approach can be generalized to accept data from remote sensing satellites. This will enable in-situ sensors to be regulated using forecasts generated by assimilating local high resolution in-situ measurements with wide-area observations from remote sensing satellites.
ACS sampling system: design, implementation, and performance evaluation
NASA Astrophysics Data System (ADS)
Di Marcantonio, Paolo; Cirami, Roberto; Chiozzi, Gianluca
2004-09-01
By means of ACS (ALMA Common Software) framework we designed and implemented a sampling system which allows sampling of every Characteristic Component Property with a specific, user-defined, sustained frequency limited only by the hardware. Collected data are sent to various clients (one or more Java plotting widgets, a dedicated GUI or a COTS application) using the ACS/CORBA Notification Channel. The data transport is optimized: samples are cached locally and sent in packets with a lower and user-defined frequency to keep network load under control. Simultaneous sampling of the Properties of different Components is also possible. Together with the design and implementation issues we present the performance of the sampling system evaluated on two different platforms: on a VME based system using VxWorks RTOS (currently adopted by ALMA) and on a PC/104+ embedded platform using Red Hat 9 Linux operating system. The PC/104+ solution offers, as an alternative, a low cost PC compatible hardware environment with free and open operating system.
The Quality System Implementation Plan (QSIP) describes the quality assurance and quality control procedures developed for the CTEPP study. It provides the QA/QC procedures used in recruitment of subjects, sample field collection, sample extraction and analysis, data storage, and...
Adaptive model predictive process control using neural networks
Buescher, K.L.; Baum, C.C.; Jones, R.D.
1997-08-19
A control system for controlling the output of at least one plant process output parameter is implemented by adaptive model predictive control using a neural network. An improved method and apparatus provides for sampling plant output and control input at a first sampling rate to provide control inputs at the fast rate. The MPC system is, however, provided with a network state vector that is constructed at a second, slower rate so that the input control values used by the MPC system are averaged over a gapped time period. Another improvement is a provision for on-line training that may include difference training, curvature training, and basis center adjustment to maintain the weights and basis centers of the neural in an updated state that can follow changes in the plant operation apart from initial off-line training data. 46 figs.
Adaptive model predictive process control using neural networks
Buescher, Kevin L.; Baum, Christopher C.; Jones, Roger D.
1997-01-01
A control system for controlling the output of at least one plant process output parameter is implemented by adaptive model predictive control using a neural network. An improved method and apparatus provides for sampling plant output and control input at a first sampling rate to provide control inputs at the fast rate. The MPC system is, however, provided with a network state vector that is constructed at a second, slower rate so that the input control values used by the MPC system are averaged over a gapped time period. Another improvement is a provision for on-line training that may include difference training, curvature training, and basis center adjustment to maintain the weights and basis centers of the neural in an updated state that can follow changes in the plant operation apart from initial off-line training data.
Molecular hydrogen absorption systems in Sloan Digital Sky Survey
NASA Astrophysics Data System (ADS)
Balashev, S. A.; Klimenko, V. V.; Ivanchik, A. V.; Varshalovich, D. A.; Petitjean, P.; Noterdaeme, P.
2014-05-01
We present a systematic search for molecular hydrogen absorption systems at high redshift in quasar spectra from the Sloan Digital Sky Survey (SDSS)-II Data Release 7 and SDSS-III Data Release 9. We have selected candidates using a modified profile fitting technique taking into account that the Lyα forest can effectively mimic H2 absorption systems at the resolution of SDSS data. To estimate the confidence level of the detections, we use two methods: a Monte Carlo sampling and an analysis of control samples. The analysis of control samples allows us to define regions of the spectral quality parameter space where H2 absorption systems can be confidently identified. We find that H2 absorption systems with column densities log NH2 > 19 can be detected in only less than 3 per cent of SDSS quasar spectra. We estimate the upper limit on the detection rate of saturated H2 absorption systems (NH2 > 19) in damped Lyα (DLA) systems to be about 7 per cent. We provide a sample of 23 confident H2 absorption system candidates that would be interesting to follow up with high-resolution spectrographs. There is a 1σ r - i colour excess and non-significant AV extinction excess in quasar spectra with an H2 candidate compared to standard DLA-bearing quasar spectra. The equivalent widths of C II, Si II and Al III (but not Fe II) absorptions associated with H2 candidate DLAs are larger compared to standard DLAs. This is probably related to a larger spread in velocity of the absorption lines in the H2-bearing sample.
Submillisecond Optical Knife-Edge Testing
NASA Technical Reports Server (NTRS)
Thurlow, P.
1983-01-01
Fast computer-controlled sampling of optical knife-edge response (KER) signal increases accuracy of optical system aberration measurement. Submicrosecond-response detectors in optical focal plane convert optical signals to electrical signals converted to digital data, sampled and feed into computer for storage and subsequent analysis. Optical data are virtually free of effects of index-of-refraction gradients.
NASA Technical Reports Server (NTRS)
Rogers, Melissa J. B.; Alexander, J. I. D.; Schoess, Jeff
1993-01-01
The Honeywell In-Space Accelerometer (HISA) system collected data in the mid-deck area of the Shuttle Columbia during the flight of STS-32, January 1990. The resulting data were to be used to investigate the response of crystal microstructure to different types of residual acceleration. The HISA is designed to detect and record transient and oscillatory accelerations. The sampling and electronics package stored averaged accelerations over two sampling periods; two sampling rates were available: 1 Hz and 50 Hz. Analysis of the HISA data followed the CMMR Acceleration Data Processing Guide, considering in-house computer modelling of a float-zone indium crystal growth experiment. Characteristic examples of HISA data showing the response to the primary reaction control system, Orbiter Maneuvering System operations, and crew treadmill activity are presented. Various orbiter structural modes are excited by these and other activities.
Lincoln, Tricia A.; Horan-Ross, Debra A.; McHale, Michael R.; Lawrence, Gregory B.
2009-01-01
The laboratory for analysis of low-ionic-strength water at the U.S. Geological Survey (USGS) Water Science Center in Troy, N.Y., analyzes samples collected by USGS projects throughout the Northeast. The laboratory's quality-assurance program is based on internal and interlaboratory quality-assurance samples and quality-control procedures that were developed to ensure proper sample collection, processing, and analysis. The quality-assurance and quality-control data were stored in the laboratory's Lab Master data-management system, which provides efficient review, compilation, and plotting of data. This report presents and discusses results of quality-assurance and quality control samples analyzed from July 2003 through June 2005. Results for the quality-control samples for 20 analytical procedures were evaluated for bias and precision. Control charts indicate that data for five of the analytical procedures were occasionally biased for either high-concentration or low-concentration samples but were within control limits; these procedures were: acid-neutralizing capacity, total monomeric aluminum, pH, silicon, and sodium. Seven of the analytical procedures were biased throughout the analysis period for the high-concentration sample, but were within control limits; these procedures were: dissolved organic carbon, chloride, nitrate (ion chromatograph), nitrite, silicon, sodium, and sulfate. The calcium and magnesium procedures were biased throughout the analysis period for the low-concentration sample, but were within control limits. The total aluminum and specific conductance procedures were biased for the high-concentration and low-concentration samples, but were within control limits. Results from the filter-blank and analytical-blank analyses indicate that the procedures for 17 of 18 analytes were within control limits, although the concentrations for blanks were occasionally outside the control limits. The data-quality objective was not met for dissolved organic carbon. Sampling and analysis precision are evaluated herein in terms of the coefficient of variation obtained for triplicate samples in the procedures for 18 of the 22 analytes. At least 85 percent of the samples met data-quality objectives for all analytes except total monomeric aluminum (82 percent of samples met objectives), total aluminum (77 percent of samples met objectives), chloride (80 percent of samples met objectives), fluoride (76 percent of samples met objectives), and nitrate (ion chromatograph) (79 percent of samples met objectives). The ammonium and total dissolved nitrogen did not meet the data-quality objectives. Results of the USGS interlaboratory Standard Reference Sample (SRS) Project indicated good data quality over the time period, with ratings for each sample in the satisfactory, good, and excellent ranges or less than 10 percent error. The P-sample (low-ionic-strength constituents) analysis had one marginal and two unsatisfactory ratings for the chloride procedure. The T-sample (trace constituents)analysis had two unsatisfactory ratings and one high range percent error for the aluminum procedure. The N-sample (nutrient constituents) analysis had one marginal rating for the nitrate procedure. Results of Environment Canada's National Water Research Institute (NWRI) program indicated that at least 84 percent of the samples met data-quality objectives for 11 of the 14 analytes; the exceptions were ammonium, total aluminum, and acid-neutralizing capacity. The ammonium procedure did not meet data quality objectives in all studies. Data-quality objectives were not met in 23 percent of samples analyzed for total aluminum and 45 percent of samples analyzed acid-neutralizing capacity. Results from blind reference-sample analyses indicated that data-quality objectives were met by at least 86 percent of the samples analyzed for calcium, chloride, fluoride, magnesium, pH, potassium, sodium, and sulfate. Data-quality objectives were not met by samples analyzed for fluoride.
Implementation guide for turbidity threshold sampling: principles, procedures, and analysis
Jack Lewis; Rand Eads
2009-01-01
Turbidity Threshold Sampling uses real-time turbidity and river stage information to automatically collect water quality samples for estimating suspended sediment loads. The system uses a programmable data logger in conjunction with a stage measurement device, a turbidity sensor, and a pumping sampler. Specialized software enables the user to control the sampling...
Time and frequency domain analysis of sampled data controllers via mixed operation equations
NASA Technical Reports Server (NTRS)
Frisch, H. P.
1981-01-01
Specification of the mathematical equations required to define the dynamic response of a linear continuous plant, subject to sampled data control, is complicated by the fact that the digital components of the control system cannot be modeled via linear ordinary differential equations. This complication can be overcome by introducing two new mathematical operations; namely, the operation of zero order hold and digial delay. It is shown that by direct utilization of these operations, a set of linear mixed operation equations can be written and used to define the dynamic response characteristics of the controlled system. It also is shown how these linear mixed operation equations lead, in an automatable manner, directly to a set of finite difference equations which are in a format compatible with follow on time and frequency domain analysis methods.
Lincoln, Tricia A.; Horan-Ross, Debra A.; McHale, Michael R.; Lawrence, Gregory B.
2009-01-01
The laboratory for analysis of low-ionic-strength water at the U.S. Geological Survey (USGS) Water Science Center in Troy, N.Y., analyzes samples collected by USGS projects throughout the Northeast. The laboratory's quality-assurance program is based on internal and interlaboratory quality-assurance samples and quality-control procedures that were developed to ensure proper sample collection, processing, and analysis. The quality-assurance and quality-control data were stored in the laboratory's Lab Master data-management system, which provides efficient review, compilation, and plotting of data. This report presents and discusses results of quality-assurance and quality control samples analyzed from July 2005 through June 2007. Results for the quality-control samples for 19 analytical procedures were evaluated for bias and precision. Control charts indicate that data for eight of the analytical procedures were occasionally biased for either high-concentration or low-concentration samples but were within control limits; these procedures were: total aluminum, calcium, magnesium, nitrate (colorimetric method), potassium, silicon, sodium, and sulfate. Eight of the analytical procedures were biased throughout the analysis period for the high-concentration sample, but were within control limits; these procedures were: total aluminum, calcium, dissolved organic carbon, chloride, nitrate (ion chromatograph), potassium, silicon, and sulfate. The magnesium and pH procedures were biased throughout the analysis period for the low-concentration sample, but were within control limits. The acid-neutralizing capacity, total monomeric aluminum, nitrite, and specific conductance procedures were biased for the high-concentration and low-concentration samples, but were within control limits. Results from the filter-blank and analytical-blank analyses indicated that the procedures for 16 of 17 analytes were within control limits, although the concentrations for blanks were occasionally outside the control limits. The data-quality objective was not met for dissolved organic carbon. Sampling and analysis precision are evaluated herein in terms of the coefficient of variation obtained for triplicate samples in the procedures for 18 of the 21 analytes. At least 93 percent of the samples met data-quality objectives for all analytes except acid-neutralizing capacity (85 percent of samples met objectives), total monomeric aluminum (83 percent of samples met objectives), total aluminum (85 percent of samples met objectives), and chloride (85 percent of samples met objectives). The ammonium and total dissolved nitrogen did not meet the data-quality objectives. Results of the USGS interlaboratory Standard Reference Sample (SRS) Project met the Troy Laboratory data-quality objectives for 87 percent of the samples analyzed. The P-sample (low-ionic-strength constituents) analysis had two outliers each in two studies. The T-sample (trace constituents) analysis and the N-sample (nutrient constituents) analysis had one outlier each in two studies. Results of Environment Canada's National Water Research Institute (NWRI) program indicated that at least 85 percent of the samples met data-quality objectives for 11 of the 14 analytes; the exceptions were acid-neutralizing capacity, total aluminum and ammonium. Data-quality objectives were not met in 41 percent of samples analyzed for acid-neutralizing capacity, 50 percent of samples analyzed for total aluminum, and 44 percent of samples analyzed for ammonium. Results from blind reference-sample analyses indicated that data-quality objectives were met by at least 86 percent of the samples analyzed for calcium, magnesium, pH, potassium, and sodium. Data-quality objectives were met by 76 percent of the samples analyzed for chloride, 80 percent of the samples analyzed for specific conductance, and 77 percent of the samples analyzed for sulfate.
Flow through electrode with automated calibration
Szecsody, James E [Richland, WA; Williams, Mark D [Richland, WA; Vermeul, Vince R [Richland, WA
2002-08-20
The present invention is an improved automated flow through electrode liquid monitoring system. The automated system has a sample inlet to a sample pump, a sample outlet from the sample pump to at least one flow through electrode with a waste port. At least one computer controls the sample pump and records data from the at least one flow through electrode for a liquid sample. The improvement relies upon (a) at least one source of a calibration sample connected to (b) an injection valve connected to said sample outlet and connected to said source, said injection valve further connected to said at least one flow through electrode, wherein said injection valve is controlled by said computer to select between said liquid sample or said calibration sample. Advantages include improved accuracy because of more frequent calibrations, no additional labor for calibration, no need to remove the flow through electrode(s), and minimal interruption of sampling.
High frequency signal acquisition and control system based on DSP+FPGA
NASA Astrophysics Data System (ADS)
Liu, Xiao-qi; Zhang, Da-zhi; Yin, Ya-dong
2017-10-01
This paper introduces a design and implementation of high frequency signal acquisition and control system based on DSP + FPGA. The system supports internal/external clock and internal/external trigger sampling. It has a maximum sampling rate of 400MBPS and has a 1.4GHz input bandwidth for the ADC. Data can be collected continuously or periodically in systems and they are stored in DDR2. At the same time, the system also supports real-time acquisition, the collected data after digital frequency conversion and Cascaded Integrator-Comb (CIC) filtering, which then be sent to the CPCI bus through the high-speed DSP, can be assigned to the fiber board for subsequent processing. The system integrates signal acquisition and pre-processing functions, which uses high-speed A/D, high-speed DSP and FPGA mixed technology and has a wide range of uses in data acquisition and recording. In the signal processing, the system can be seamlessly connected to the dedicated processor board. The system has the advantages of multi-selectivity, good scalability and so on, which satisfies the different requirements of different signals in different projects.
Application of higher harmonic blade feathering for helicopter vibration reduction
NASA Technical Reports Server (NTRS)
Powers, R. W.
1978-01-01
Higher harmonic blade feathering for helicopter vibration reduction is considered. Recent wind tunnel tests confirmed the effectiveness of higher harmonic control in reducing articulated rotor vibratory hub loads. Several predictive analyses developed in support of the NASA program were shown to be capable of calculating single harmonic control inputs required to minimize a single 4P hub response. In addition, a multiple-input, multiple-output harmonic control predictive analysis was developed. All techniques developed thus far obtain a solution by extracting empirical transfer functions from sampled data. Algorithm data sampling and processing requirements are minimal to encourage adaptive control system application of such techniques in a flight environment.
Practical continuous-variable quantum key distribution without finite sampling bandwidth effects.
Li, Huasheng; Wang, Chao; Huang, Peng; Huang, Duan; Wang, Tao; Zeng, Guihua
2016-09-05
In a practical continuous-variable quantum key distribution system, finite sampling bandwidth of the employed analog-to-digital converter at the receiver's side may lead to inaccurate results of pulse peak sampling. Then, errors in the parameters estimation resulted. Subsequently, the system performance decreases and security loopholes are exposed to eavesdroppers. In this paper, we propose a novel data acquisition scheme which consists of two parts, i.e., a dynamic delay adjusting module and a statistical power feedback-control algorithm. The proposed scheme may improve dramatically the data acquisition precision of pulse peak sampling and remove the finite sampling bandwidth effects. Moreover, the optimal peak sampling position of a pulse signal can be dynamically calibrated through monitoring the change of the statistical power of the sampled data in the proposed scheme. This helps to resist against some practical attacks, such as the well-known local oscillator calibration attack.
Lincoln, Tricia A.; Horan-Ross, Debra A.; McHale, Michael R.; Lawrence, Gregory B.
2009-01-01
The laboratory for analysis of low-ionic-strength water at the U.S. Geological Survey (USGS) Water Science Center in Troy, N.Y., analyzes samples collected by USGS projects throughout the Northeast. The laboratory's quality-assurance program is based on internal and interlaboratory quality-assurance samples and quality-control procedures that were developed to ensure proper sample collection, processing, and analysis. The quality-assurance and quality-control data were stored in the laboratory's Lab Master data-management system, which provides efficient review, compilation, and plotting of data. This report presents and discusses results of quality-assurance and quality control samples analyzed from July 2001 through June 2003. Results for the quality-control samples for 19 analytical procedures were evaluated for bias and precision. Control charts indicate that data for six of the analytical procedures were occasionally biased for either high-concentration or low-concentration samples but were within control limits; these procedures were: acid-neutralizing capacity, chloride, magnesium, nitrate (ion chromatography), potassium, and sodium. The calcium procedure was biased throughout the analysis period for the high-concentration sample, but was within control limits. The total monomeric aluminum and fluoride procedures were biased throughout the analysis period for the low-concentration sample, but were within control limits. The total aluminum, pH, specific conductance, and sulfate procedures were biased for the high-concentration and low-concentration samples, but were within control limits. Results from the filter-blank and analytical-blank analyses indicate that the procedures for 16 of 18 analytes were within control limits, although the concentrations for blanks were occasionally outside the control limits. The data-quality objective was not met for the dissolved organic carbon or specific conductance procedures. Sampling and analysis precision are evaluated herein in terms of the coefficient of variation obtained for triplicate samples in the procedures for 18 of the 21 analytes. At least 90 percent of the samples met data-quality objectives for all procedures except total monomeric aluminum (83 percent of samples met objectives), total aluminum (76 percent of samples met objectives), ammonium (73 percent of samples met objectives), dissolved organic carbon (86 percent of samples met objectives), and nitrate (81 percent of samples met objectives). The data-quality objective was not met for the nitrite procedure. Results of the USGS interlaboratory Standard Reference Sample (SRS) Project indicated satisfactory or above data quality over the time period, with most performance ratings for each sample in the good-to-excellent range. The N-sample (nutrient constituents) analysis had one unsatisfactory rating for the ammonium procedure in one study. The T-sample (trace constituents) analysis had one unsatisfactory rating for the magnesium procedure and one marginal rating for the potassium procedure in one study and one unsatisfactory rating for the sodium procedure in another. Results of Environment Canada's National Water Research Institute (NWRI) program indicated that at least 90 percent of the samples met data-quality objectives for 10 of the 14 analytes; the exceptions were acid-neutralizing capacity, ammonium, dissolved organic carbon, and sodium. Data-quality objectives were not met in 37 percent of samples analyzed for acid-neutralizing capacity, 28 percent of samples analyzed for dissolved organic carbon, and 30 percent of samples analyzed for sodium. Results indicate a positive bias for the ammonium procedure in one study and a negative bias in another. Results from blind reference-sample analyses indicated that data-quality objectives were met by at least 90 percent of the samples analyzed for calcium, chloride, magnesium, pH, potassium, and sodium. Data-quality objectives were met by 78 percent of
Compact drilling and sample system
NASA Technical Reports Server (NTRS)
Gillis-Smith, Greg R.; Petercsak, Doug
1998-01-01
The Compact Drilling and Sample System (CDSS) was developed to drill into terrestrial, cometary, and asteroid material in a cryogenic, vacuum environment in order to acquire subsurface samples. Although drills were used by the Apollo astronauts some 20 years ago, this drill is a fraction of the mass and power and operates completely autonomously, able to drill, acquire, transport, dock, and release sample containers in science instruments. The CDSS has incorporated into its control system the ability to gather science data about the material being drilled by measuring drilling rate per force applied and torque. This drill will be able to optimize rotation and thrust in order to achieve the highest drilling rate possible in any given sample. The drill can be commanded to drill at a specified force, so that force imparted on the rover or lander is limited. This paper will discuss the cryo dc brush motors, carbide gears, cryogenic lubrication, quick-release interchangeable sampling drill bits, percussion drilling and the control system developed to achieve autonomous, cryogenic, vacuum, lightweight drilling.
Chicken barn climate and hazardous volatile compounds control using simple linear regression and PID
NASA Astrophysics Data System (ADS)
Abdullah, A. H.; Bakar, M. A. A.; Shukor, S. A. A.; Saad, F. S. A.; Kamis, M. S.; Mustafa, M. H.; Khalid, N. S.
2016-07-01
The hazardous volatile compounds from chicken manure in chicken barn are potentially to be a health threat to the farm animals and workers. Ammonia (NH3) and hydrogen sulphide (H2S) produced in chicken barn are influenced by climate changes. The Electronic Nose (e-nose) is used for the barn's air, temperature and humidity data sampling. Simple Linear Regression is used to identify the correlation between temperature-humidity, humidity-ammonia and ammonia-hydrogen sulphide. MATLAB Simulink software was used for the sample data analysis using PID controller. Results shows that the performance of PID controller using the Ziegler-Nichols technique can improve the system controller to control climate in chicken barn.
Turbidity-controlled sampling for suspended sediment load estimation
Jack Lewis
2003-01-01
Abstract - Automated data collection is essential to effectively measure suspended sediment loads in storm events, particularly in small basins. Continuous turbidity measurements can be used, along with discharge, in an automated system that makes real-time sampling decisions to facilitate sediment load estimation. The Turbidity Threshold Sampling method distributes...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tandon, Lav; Colletti, Lisa M.; Drake, Lawrence R.
This report discusses the process used to prove in the SRNL-Rev.2 coulometer for isotopic data analysis used in the special plutonium material project. In May of 2012, the PAR 173 coulometer system that had been the workhorse of the Plutonium Assay team since the early 1970s became inoperable. A new coulometer system had been purchased from Savannah River National Laboratory (SRNL) and installed in August of 2011. Due to funding issues the new system was not qualified at that time. Following the failure of the PAR 173, it became necessary to qualify the new system for use in Process 3401a,more » Plutonium Assay by Controlled Coulometry. A qualification plan similar to what is described in PQR -141a was followed. Experiments were performed to establish a statistical summary of the performance of the new system by monitoring the repetitive analysis of quality control sample, PEOL, and the assay of plutonium metals obtained from the Plutonium Exchange Program. The data for the experiments was acquired using work instructions ANC125 and ANC195. Figure 1 shows approximately 2 years of data for the PEOL material obtained using the PAR 173. The required acceptance criteria for the sample are that it returns the correct value for the quality control material of 88.00% within 2 sigma (95% Confidence Interval). It also must meet daily precision standards that are set from the historical data analysis of decades of data. The 2 sigma value that is currently used is 0.146 % as evaluated by the Statistical Science Group, CCS-6. The average value of the PEOL quality control material run in 10 separate days on the SRNL-03 coulometer is 87.98% with a relative standard deviation of 0.04 at the 95% Confidence interval. The date of data acquisition is between 5/23/2012 to 8/1/2012. The control samples are run every day experiments using the coulometer are carried out. It is also used to prove an instrument is in statistical control before any experiments are undertaken. The total number of replicate controls run with the new coulometer to date, is n=18. This value is identical to that calculated by the LANL statistical group for this material from data produced by the PAR 173 system over the period of October 2007 to May 2011. The final validation/verification test was to run a blind sample over multiple days. AAC participates in a plutonium exchange program which supplies blind Pu metal samples to the group on a regular basis. The Pu material supplied for this study was ran using the PAR 173 in the past and more recently with the new system. Table 1a contains the values determined through the use of the PAR 173 and Table 1b contains the values obtained with the new system. The Pu assay value obtained on the SRNL system is for paired analysis and had a value of 98.88+/-0.07% RSD at 95% CI. The Pu assay value (decay corrected to July 2012) of the material determined in prior measurements using the PAR173 is 99.05 +/- 0.06 % RSD at 95% CI. We believe that the instrument is adequate to meet the needs of the program.« less
Converging Redundant Sensor Network Information for Improved Building Control
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dale Tiller; D. Phil; Gregor Henze
2007-09-30
This project investigated the development and application of sensor networks to enhance building energy management and security. Commercial, industrial and residential buildings often incorporate systems used to determine occupancy, but current sensor technology and control algorithms limit the effectiveness of these systems. For example, most of these systems rely on single monitoring points to detect occupancy, when more than one monitoring point could improve system performance. Phase I of the project focused on instrumentation and data collection. During the initial project phase, a new occupancy detection system was developed, commissioned and installed in a sample of private offices and open-planmore » office workstations. Data acquisition systems were developed and deployed to collect data on space occupancy profiles. Phase II of the project demonstrated that a network of several sensors provides a more accurate measure of occupancy than is possible using systems based on single monitoring points. This phase also established that analysis algorithms could be applied to the sensor network data stream to improve the accuracy of system performance in energy management and security applications. In Phase III of the project, the sensor network from Phase I was complemented by a control strategy developed based on the results from the first two project phases: this controller was implemented in a small sample of work areas, and applied to lighting control. Two additional technologies were developed in the course of completing the project. A prototype web-based display that portrays the current status of each detector in a sensor network monitoring building occupancy was designed and implemented. A new capability that enables occupancy sensors in a sensor network to dynamically set the 'time delay' interval based on ongoing occupant behavior in the space was also designed and implemented.« less
Development of Portable, Wireless and Smartphone Controllable Near-Infrared Spectroscopy System.
Watanabe, Takashi; Sekine, Rui; Mizuno, Toshihiko; Miwa, Mitsuharu
We have developed portable near-infrared tissue oxygenation monitoring systems, called the "PocketNIRS Duo" and the "PocketNIRS HM", which features wireless data communication and a sampling rate of up to 60 data readings per second. The systems can be controlled by smartphone or personal computer. We demonstrate the efficacy of the systems for monitoring changes in brain and arm muscle hemodynamics and oxygenation in breath-holding and cuff-occlusion tests, respectively.Our systems should prove to be useful as an oxygenation monitor not only in research but also in healthcare applications.
System identification from closed-loop data with known output feedback dynamics
NASA Technical Reports Server (NTRS)
Phan, Minh; Juang, Jer-Nan; Horta, Lucas G.; Longman, Richard W.
1992-01-01
This paper presents a procedure to identify the open loop systems when it is operating under closed loop conditions. First, closed loop excitation data are used to compute the system open loop and closed loop Markov parameters. The Markov parameters, which are the pulse response samples, are then used to compute a state space representation of the open loop system. Two closed loop configurations are considered in this paper. The closed loop system can have either a linear output feedback controller or a dynamic output feedback controller. Numerical examples are provided to illustrate the proposed closed loop identification method.
NASA Technical Reports Server (NTRS)
Fabiniak, R. C.; Fabiniak, T. J.
1971-01-01
The results of experiments 1, 2, and 10 of the Apollo 14 composite casting demonstration are discussed. The purpose of the demonstration, with regard to samples 1 and 2, was to obtain preliminary data on the liquid phase sintering process in a weightless environment. With regard to sample 10, the purpose was to obtain preliminary information on how to achieve uniform dispersion of dense particles on a metal matrix by employing shaking modes or forces in the system when the metal matrix is molten. Results of the demonstrations were interpreted in a quantitative and qualitative manner. For experiment 1 it was found that the tungsten particles were redistributed more uniformly in the flight sample than in the control sample. Experiment 2 results indicate that complete melting may not have occured and thus a high degree of significance cannot be associated with the qualitative results relating to particle redistribution data. The particle-matrix system of experiment 10 was found to be nonwetting.
Dharani, S; Rakkiyappan, R; Cao, Jinde; Alsaedi, Ahmed
2017-08-01
This paper explores the problem of synchronization of a class of generalized reaction-diffusion neural networks with mixed time-varying delays. The mixed time-varying delays under consideration comprise of both discrete and distributed delays. Due to the development and merits of digital controllers, sampled-data control is a natural choice to establish synchronization in continuous-time systems. Using a newly introduced integral inequality, less conservative synchronization criteria that assure the global asymptotic synchronization of the considered generalized reaction-diffusion neural network and mixed delays are established in terms of linear matrix inequalities (LMIs). The obtained easy-to-test LMI-based synchronization criteria depends on the delay bounds in addition to the reaction-diffusion terms, which is more practicable. Upon solving these LMIs by using Matlab LMI control toolbox, a desired sampled-data controller gain can be acuqired without any difficulty. Finally, numerical examples are exploited to express the validity of the derived LMI-based synchronization criteria.
A system architecture for online data interpretation and reduction in fluorescence microscopy
NASA Astrophysics Data System (ADS)
Röder, Thorsten; Geisbauer, Matthias; Chen, Yang; Knoll, Alois; Uhl, Rainer
2010-01-01
In this paper we present a high-throughput sample screening system that enables real-time data analysis and reduction for live cell analysis using fluorescence microscopy. We propose a novel system architecture capable of analyzing a large amount of samples during the experiment and thus greatly minimizing the post-analysis phase that is the common practice today. By utilizing data reduction algorithms, relevant information of the target cells is extracted from the online collected data stream, and then used to adjust the experiment parameters in real-time, allowing the system to dynamically react on changing sample properties and to control the microscope setup accordingly. The proposed system consists of an integrated DSP-FPGA hybrid solution to ensure the required real-time constraints, to execute efficiently the underlying computer vision algorithms and to close the perception-action loop. We demonstrate our approach by addressing the selective imaging of cells with a particular combination of markers. With this novel closed-loop system the amount of superfluous collected data is minimized, while at the same time the information entropy increases.
Data-Based Predictive Control with Multirate Prediction Step
NASA Technical Reports Server (NTRS)
Barlow, Jonathan S.
2010-01-01
Data-based predictive control is an emerging control method that stems from Model Predictive Control (MPC). MPC computes current control action based on a prediction of the system output a number of time steps into the future and is generally derived from a known model of the system. Data-based predictive control has the advantage of deriving predictive models and controller gains from input-output data. Thus, a controller can be designed from the outputs of complex simulation code or a physical system where no explicit model exists. If the output data happens to be corrupted by periodic disturbances, the designed controller will also have the built-in ability to reject these disturbances without the need to know them. When data-based predictive control is implemented online, it becomes a version of adaptive control. One challenge of MPC is computational requirements increasing with prediction horizon length. This paper develops a closed-loop dynamic output feedback controller that minimizes a multi-step-ahead receding-horizon cost function with multirate prediction step. One result is a reduced influence of prediction horizon and the number of system outputs on the computational requirements of the controller. Another result is an emphasis on portions of the prediction window that are sampled more frequently. A third result is the ability to include more outputs in the feedback path than in the cost function.
MICROPROCESSOR-BASED DATA-ACQUISITION SYSTEM FOR A BOREHOLE RADAR.
Bradley, Jerry A.; Wright, David L.
1987-01-01
An efficient microprocessor-based system is described that permits real-time acquisition, stacking, and digital recording of data generated by a borehole radar system. Although the system digitizes, stacks, and records independently of a computer, it is interfaced to a desktop computer for program control over system parameters such as sampling interval, number of samples, number of times the data are stacked prior to recording on nine-track tape, and for graphics display of the digitized data. The data can be transferred to the desktop computer during recording, or it can be played back from a tape at a latter time. Using the desktop computer, the operator observes results while recording data and generates hard-copy graphics in the field. Thus, the radar operator can immediately evaluate the quality of data being obtained, modify system parameters, study the radar logs before leaving the field, and rerun borehole logs if necessary. The system has proven to be reliable in the field and has increased productivity both in the field and in the laboratory.
NASA Astrophysics Data System (ADS)
Fitzgerald, Ryan; Karanassios, Vassili
2017-05-01
There are many applications requiring chemical analysis in the field and analytical results in (near) real-time. For example, when accidental spills occur. In others, collecting samples in the field followed by analysis in a lab increases costs and introduces time-delays. In such cases, "bring part of the lab to the sample" would be ideal. Toward this ideal (and to further reduce size and weight), we developed a relatively inexpensive, battery-operated, wireless data acquisition hardware system around an Arduino nano micro-controller and a 16-bit ADC (Analog-to- Digital Converter) with a max sampling rate of 860 samples/s. The hardware communicates the acquired data using low-power Bluetooth. Software for data acquisition and data display was written in Python. Potential ways of making the hardware-software approach described here a part of the Internet-of-Things (IoT) are presented.
Oden, Timothy D.
2011-01-01
The Gulf Coast aquifer system is the primary water supply for Montgomery County in southeastern Texas, including part of the Houston metropolitan area and the cities of Magnolia, Conroe, and The Woodlands Township, Texas. The U.S. Geological Survey, in cooperation with the Lone Star Groundwater Conservation District, collected environmental tracer data in the Gulf Coast aquifer system, primarily in Montgomery County. Forty existing groundwater wells screened in the Gulf Coast aquifer system were selected for sampling in Montgomery County (38 wells), Waller County (1 well), and Walker County (1 well). Groundwater-quality samples, physicochemical properties, and water-level data were collected once from each of the 40 wells during March-September 2008. Groundwater-quality samples were analyzed for dissolved gases and the environmental tracers sulfur hexafluoride, chlorofluorocarbons, tritium, helium-4, and helium-3/tritium. Water samples were collected and processed onsite using methods designed to minimize changes to the water-sample chemistry or contamination from the atmosphere. Replicate samples for quality assurance and quality control were collected with each environmental sample. Well-construction information and environmental tracer data for March-September 2008 are presented.
Digital redesign of the control system for the Robotics Research Corporation model K-1607 robot
NASA Technical Reports Server (NTRS)
Carroll, Robert L.
1989-01-01
The analog control system for positioning each link of the Robotics Research Corporation Model K-1607 robot manipulator was redesigned for computer control. In order to accomplish the redesign, a linearized model of the dynamic behavior of the robot was developed. The parameters of the model were determined by examination of the input-output data collected in closed-loop operation of the analog control system. The robot manipulator possesses seven degrees of freedom in its motion. The analog control system installed by the manufacturer of the robot attempts to control the positioning of each link without feedback from other links. Constraints on the design of a digital control system include: the robot cannot be disassembled for measurement of parameters; the digital control system must not include filtering operations if possible, because of lack of computer capability; and criteria of goodness of control system performing is lacking. The resulting design employs sampled-data position and velocity feedback. The criteria of the design permits the control system gain margin and phase margin, measured at the same frequencies, to be the same as that provided by the analog control system.
Wireless sensing and vibration control with increased redundancy and robustness design.
Li, Peng; Li, Luyu; Song, Gangbing; Yu, Yan
2014-11-01
Control systems with long distance sensor and actuator wiring have the problem of high system cost and increased sensor noise. Wireless sensor network (WSN)-based control systems are an alternative solution involving lower setup and maintenance costs and reduced sensor noise. However, WSN-based control systems also encounter problems such as possible data loss, irregular sampling periods (due to the uncertainty of the wireless channel), and the possibility of sensor breakdown (due to the increased complexity of the overall control system). In this paper, a wireless microcontroller-based control system is designed and implemented to wirelessly perform vibration control. The wireless microcontroller-based system is quite different from regular control systems due to its limited speed and computational power. Hardware, software, and control algorithm design are described in detail to demonstrate this prototype. Model and system state compensation is used in the wireless control system to solve the problems of data loss and sensor breakdown. A positive position feedback controller is used as the control law for the task of active vibration suppression. Both wired and wireless controllers are implemented. The results show that the WSN-based control system can be successfully used to suppress the vibration and produces resilient results in the presence of sensor failure.
The Design of Feedback Control Systems Containing a Saturation Type Nonlinearity
NASA Technical Reports Server (NTRS)
Schmidt, Stanley F.; Harper, Eleanor V.
1960-01-01
A derivation of the optimum response for a step input for plant transfer functions which have an unstable pole and further data on plants with a single zero in the left half of the s plane. The calculated data are presented tabulated in normalized form. Optimum control systems are considered. The optimum system is defined as one which keeps the error as small as possible regardless of the input, under the constraint that the input to the plant (or controlled system) is limited. Intuitive arguments show that in the case where only the error can be sensed directly, the optimum system is obtained from the optimum relay or on-off solution. References to known solutions are presented. For the case when the system is of the sampled-data type, arguments are presented which indicate the optimum sampled-data system may be extremely difficult if not impossible to realize practically except for very simple plant transfer functions. Two examples of aircraft attitude autopilots are presented, one for a statically stable and the other for a statically unstable airframe. The rate of change of elevator motion is assumed limited for these examples. It is shown that by use of nonlinear design techniques described in NASA TN D-20 one can obtain near optimum response for step inputs and reason- able response to sine wave inputs for either case. Also, the nonlinear design prevents inputs from driving the system unstable for either case.
Portable Dew Point Mass Spectrometry System for Real-Time Gas and Moisture Analysis
NASA Technical Reports Server (NTRS)
Arkin, C.; Gillespie, Stacey; Ratzel, Christopher
2010-01-01
A portable instrument incorporates both mass spectrometry and dew point measurement to provide real-time, quantitative gas measurements of helium, nitrogen, oxygen, argon, and carbon dioxide, along with real-time, quantitative moisture analysis. The Portable Dew Point Mass Spectrometry (PDP-MS) system comprises a single quadrupole mass spectrometer and a high vacuum system consisting of a turbopump and a diaphragm-backing pump. A capacitive membrane dew point sensor was placed upstream of the MS, but still within the pressure-flow control pneumatic region. Pressure-flow control was achieved with an upstream precision metering valve, a capacitance diaphragm gauge, and a downstream mass flow controller. User configurable LabVIEW software was developed to provide real-time concentration data for the MS, dew point monitor, and sample delivery system pressure control, pressure and flow monitoring, and recording. The system has been designed to include in situ, NIST-traceable calibration. Certain sample tubing retains sufficient water that even if the sample is dry, the sample tube will desorb water to an amount resulting in moisture concentration errors up to 500 ppm for as long as 10 minutes. It was determined that Bev-A-Line IV was the best sample line to use. As a result of this issue, it is prudent to add a high-level humidity sensor to PDP-MS so such events can be prevented in the future.
Electronic integrated disease surveillance system and pathogen asset control system.
Wahl, Tom G; Burdakov, Aleksey V; Oukharov, Andrey O; Zhilokov, Azamat K
2012-06-20
Electronic Integrated Disease Surveillance System (EIDSS) has been used to strengthen and support monitoring and prevention of dangerous diseases within One Health concept by integrating veterinary and human surveillance, passive and active approaches, case-based records including disease-specific clinical data based on standardised case definitions and aggregated data, laboratory data including sample tracking linked to each case and event with test results and epidemiological investigations. Information was collected and shared in secure way by different means: through the distributed nodes which are continuously synchronised amongst each other, through the web service, through the handheld devices. Electronic Integrated Disease Surveillance System provided near real time information flow that has been then disseminated to the appropriate organisations in a timely manner. It has been used for comprehensive analysis and visualisation capabilities including real time mapping of case events as these unfold enhancing decision making. Electronic Integrated Disease Surveillance System facilitated countries to comply with the IHR 2005 requirements through a data transfer module reporting diseases electronically to the World Health Organisation (WHO) data center as well as establish authorised data exchange with other electronic system using Open Architecture approach. Pathogen Asset Control System (PACS) has been used for accounting, management and control of biological agent stocks. Information on samples and strains of any kind throughout their entire lifecycle has been tracked in a comprehensive and flexible solution PACS.Both systems have been used in a combination and individually. Electronic Integrated Disease Surveillance System and PACS are currently deployed in the Republics of Kazakhstan, Georgia and Azerbaijan as a part of the Cooperative Biological Engagement Program (CBEP) sponsored by the US Defense Threat Reduction Agency (DTRA).
Weng, Falu; Liu, Mingxin; Mao, Weijie; Ding, Yuanchun; Liu, Feifei
2018-05-10
The problem of sampled-data-based vibration control for structural systems with finite-time state constraint and sensor outage is investigated in this paper. The objective of designing controllers is to guarantee the stability and anti-disturbance performance of the closed-loop systems while some sensor outages happen. Firstly, based on matrix transformation, the state-space model of structural systems with sensor outages and uncertainties appearing in the mass, damping and stiffness matrices is established. Secondly, by considering most of those earthquakes or strong winds happen in a very short time, and it is often the peak values make the structures damaged, the finite-time stability analysis method is introduced to constrain the state responses in a given time interval, and the H-infinity stability is adopted in the controller design to make sure that the closed-loop system has a prescribed level of disturbance attenuation performance during the whole control process. Furthermore, all stabilization conditions are expressed in the forms of linear matrix inequalities (LMIs), whose feasibility can be easily checked by using the LMI Toolbox. Finally, numerical examples are given to demonstrate the effectiveness of the proposed theorems. Copyright © 2018 ISA. Published by Elsevier Ltd. All rights reserved.
Quality-control materials in the USDA National Food and Nutrient Analysis Program (NFNAP).
Phillips, Katherine M; Patterson, Kristine Y; Rasor, Amy S; Exler, Jacob; Haytowitz, David B; Holden, Joanne M; Pehrsson, Pamela R
2006-03-01
The US Department of Agriculture (USDA) Nutrient Data Laboratory (NDL) develops and maintains the USDA National Nutrient Databank System (NDBS). Data are released from the NDBS for scientific and public use through the USDA National Nutrient Database for Standard Reference (SR) ( http://www.ars.usda.gov/ba/bhnrc/ndl ). In 1997 the NDL initiated the National Food and Nutrient Analysis Program (NFNAP) to update and expand its food-composition data. The program included: 1) nationwide probability-based sampling of foods; 2) central processing and archiving of food samples; 3) analysis of food components at commercial, government, and university laboratories; 4) incorporation of new analytical data into the NDBS; and 5) dissemination of these data to the scientific community. A key feature and strength of the NFNAP was a rigorous quality-control program that enabled independent verification of the accuracy and precision of analytical results. Custom-made food-control composites and/or commercially available certified reference materials were sent to the laboratories, blinded, with the samples. Data for these materials were essential to ongoing monitoring of analytical work, to identify and resolve suspected analytical problems, to ensure the accuracy and precision of results for the NFNAP food samples.
NASA Astrophysics Data System (ADS)
Belapurkar, Rohit K.
Future aircraft engine control systems will be based on a distributed architecture, in which, the sensors and actuators will be connected to the Full Authority Digital Engine Control (FADEC) through an engine area network. Distributed engine control architecture will allow the implementation of advanced, active control techniques along with achieving weight reduction, improvement in performance and lower life cycle cost. The performance of a distributed engine control system is predominantly dependent on the performance of the communication network. Due to the serial data transmission policy, network-induced time delays and sampling jitter are introduced between the sensor/actuator nodes and the distributed FADEC. Communication network faults and transient node failures may result in data dropouts, which may not only degrade the control system performance but may even destabilize the engine control system. Three different architectures for a turbine engine control system based on a distributed framework are presented. A partially distributed control system for a turbo-shaft engine is designed based on ARINC 825 communication protocol. Stability conditions and control design methodology are developed for the proposed partially distributed turbo-shaft engine control system to guarantee the desired performance under the presence of network-induced time delay and random data loss due to transient sensor/actuator failures. A fault tolerant control design methodology is proposed to benefit from the availability of an additional system bandwidth and from the broadcast feature of the data network. It is shown that a reconfigurable fault tolerant control design can help to reduce the performance degradation in presence of node failures. A T-700 turbo-shaft engine model is used to validate the proposed control methodology based on both single input and multiple-input multiple-output control design techniques.
Systematic plan of building Web geographic information system based on ActiveX control
NASA Astrophysics Data System (ADS)
Zhang, Xia; Li, Deren; Zhu, Xinyan; Chen, Nengcheng
2003-03-01
A systematic plan of building Web Geographic Information System (WebGIS) using ActiveX technology is proposed in this paper. In the proposed plan, ActiveX control technology is adopted in building client-side application, and two different schemas are introduced to implement communication between controls in users¡ browser and middle application server. One is based on Distribute Component Object Model (DCOM), the other is based on socket. In the former schema, middle service application is developed as a DCOM object that communicates with ActiveX control through Object Remote Procedure Call (ORPC) and accesses data in GIS Data Server through Open Database Connectivity (ODBC). In the latter, middle service application is developed using Java language. It communicates with ActiveX control through socket based on TCP/IP and accesses data in GIS Data Server through Java Database Connectivity (JDBC). The first one is usually developed using C/C++, and it is difficult to develop and deploy. The second one is relatively easy to develop, but its performance of data transfer relies on Web bandwidth. A sample application is developed using the latter schema. It is proved that the performance of the sample application is better than that of some other WebGIS applications in some degree.
Nonlinear discrete-time multirate adaptive control of non-linear vibrations of smart beams
NASA Astrophysics Data System (ADS)
Georgiou, Georgios; Foutsitzi, Georgia A.; Stavroulakis, Georgios E.
2018-06-01
The nonlinear adaptive digital control of a smart piezoelectric beam is considered. It is shown that in the case of a sampled-data context, a multirate control strategy provides an appropriate framework in order to achieve vibration regulation, ensuring the stability of the whole control system. Under parametric uncertainties in the model parameters (damping ratios, frequencies, levels of non linearities and cross coupling, control input parameters), the scheme is completed with an adaptation law deduced from hyperstability concepts. This results in the asymptotic satisfaction of the control objectives at the sampling instants. Simulation results are presented.
Dynamics of multirate sampled data control systems. [for space shuttle boost vehicle
NASA Technical Reports Server (NTRS)
Naylor, J. R.; Hynes, R. J.; Molnar, D. O.
1974-01-01
The effect was investigated of the synthesis approach (single or multirate) on the machine requirements for a digital control system for the space shuttle boost vehicle. The study encompassed four major work areas: synthesis approach trades, machine requirements trades, design analysis requirements and multirate adaptive control techniques. The primary results are two multirate autopilot designs for the low Q and maximum Q flight conditions that exhibits equal or better performance than the analog and single rate system designs. Also, a preferred technique for analyzing and synthesizing multirate digital control systems is included.
ICS-II USA research design and methodology.
Rana, H; Andersen, R M; Nakazono, T T; Davidson, P L
1997-05-01
The purpose of the WHO-sponsored International Collaborative Study of Oral Health Outcomes (ICS-II) was to provide policy-markers and researchers with detailed, reliable, and valid data on the oral health situation in their countries or regions, together with comparative data from other dental care delivery systems. ICS-II used a cross-sectional design with no explicit control groups or experimental interventions. A standardized methodology was developed and tested for collecting and analyzing epidemiological, sociocultural, economic, and delivery system data. Respondent information was obtained by household interviews, and clinical examinations were conducted by calibrated oral epidemiologists. Discussed are the sampling design characteristics for the USA research locations, response rates, samples size for interview and oral examination data, weighting procedures, and statistical methods. SUDAAN was used to adjust variance calculations, since complex sampling designs were used.
NASA Astrophysics Data System (ADS)
Klein, Matthias; Vaes, W. H. J.; Fabriek, B.; Sandman, H.; Mous, D. J. W.; Gottdang, A.
2013-01-01
The Netherlands Organization for Applied Scientific Research (TNO) has installed a compact 1 MV multi-element AMS system manufactured by High Voltage Engineering Europa B.V., The Netherlands. TNO performs clinical research programs for pharmaceutical and innovative foods industry to obtain early pharmacokinetic data and to provide anti-osteoporotic efficacy data of new treatments. The AMS system will analyze carbon, iodine and calcium samples for this purpose. The first measurements on blank samples indicate background levels in the low 10-12 for calcium and iodine, making the system well suited for these biomedical applications. Carbon blanks have been measured at low 10-16. For unattended, around-the-clock analysis, the system features the 200 sample version of the SO110 hybrid ion source and user friendly control software.
NASA Astrophysics Data System (ADS)
Ryan, D. P.; Roth, G. S.
1982-04-01
Complete documentation of the 15 programs and 11 data files of the EPA Atomic Absorption Instrument Automation System is presented. The system incorporates the following major features: (1) multipoint calibration using first, second, or third degree regression or linear interpolation, (2) timely quality control assessments for spiked samples, duplicates, laboratory control standards, reagent blanks, and instrument check standards, (3) reagent blank subtraction, and (4) plotting of calibration curves and raw data peaks. The programs of this system are written in Data General Extended BASIC, Revision 4.3, as enhanced for multi-user, real-time data acquisition. They run in a Data General Nova 840 minicomputer under the operating system RDOS, Revision 6.2. There is a functional description, a symbol definitions table, a functional flowchart, a program listing, and a symbol cross reference table for each program. The structure of every data file is also detailed.
General purpose rocket furnace
NASA Technical Reports Server (NTRS)
Aldrich, B. R.; Whitt, W. D. (Inventor)
1979-01-01
A multipurpose furnace for space vehicles used for material processing experiments in an outer space environment is described. The furnace contains three separate cavities designed to process samples of the widest possible range of materials and thermal requirements. Each cavity contains three heating elements capable of independent function under the direction of an automatic and programmable control system. A heat removable mechanism is also provided for each cavity which operates in conjunction with the control system for establishing an isothermally heated cavity or a wide range of thermal gradients and cool down rates. A monitoring system compatible with the rocket telemetry provides furnace performance and sample growth rate data throughout the processing cycle.
Instrumentation for a dry-pond detention study
Pope, L.M.; Jennings, M.E.; Thibodeaux, K.G.
1988-01-01
A 12.3-acre, fully urbanized, residential land-use catchment was instrumented by the U. S. Geological Survey in Topeka, Kansas. Hydraulic instrumentation for flow measurement includes two types of flumes, a pipe-insert flume and a culvert-inlet (manhole) flume. Samples of rainfall and runoff for water-quality analyses were collected by automatic, 3-liter, 24-sample capacity water samples controlled by multichannel data loggers. Ancillary equipment included a raingage and wet/dry atmospheric-deposition sampler. Nineteen stormwater runoff events were monitored at the site using the instrumentation system. The system has a high reliability of data capture and permits an accurate determination of storm-water loads.
A demonstration of NIOSH push-pull ventilation criteria.
Klein, M K
1987-03-01
This paper summarizes the results of a study performed on an actual chrome plating tank in order to validate criteria for push-pull ventilation systems developed by Huebener and Hughes at NIOSH. Validation of the criteria was made by taking area industrial hygiene samples for hexavalent and total chrome at ten locations around the plating tank. The sampling was performed during actual production runs or while the tank was operating with a dummy load. The sampling data are summarized. The data show that the push-pull system, operating at Huebener's criteria, could control emissions to below the current standards and guidelines. Conclusions and recommendations are included.
Event-triggered synchronization for reaction-diffusion complex networks via random sampling
NASA Astrophysics Data System (ADS)
Dong, Tao; Wang, Aijuan; Zhu, Huiyun; Liao, Xiaofeng
2018-04-01
In this paper, the synchronization problem of the reaction-diffusion complex networks (RDCNs) with Dirichlet boundary conditions is considered, where the data is sampled randomly. An event-triggered controller based on the sampled data is proposed, which can reduce the number of controller and the communication load. Under this strategy, the synchronization problem of the diffusion complex network is equivalently converted to the stability of a of reaction-diffusion complex dynamical systems with time delay. By using the matrix inequality technique and Lyapunov method, the synchronization conditions of the RDCNs are derived, which are dependent on the diffusion term. Moreover, it is found the proposed control strategy can get rid of the Zeno behavior naturally. Finally, a numerical example is given to verify the obtained results.
Investigation to develop a multistage forest sampling inventory system using ERTS-1 imagery
NASA Technical Reports Server (NTRS)
Langley, P. G.; Vanroessel, J. W. (Principal Investigator); Wert, S. L.
1975-01-01
The author has identified the following significant results. The annotation system produced a RMSE of about 200 m ground distance in the MSS data system with the control data used. All the analytical MSS interpretation models tried were highly significant. However, the gains in forest sampling efficiency that can be achieved by using the models vary from zero to over 50 percent depending on the area to which they are applied and the sampling method used. Among the sampling methods tried, regression sampling yielded substantial and the most consistent gains. The single most significant variable in the interpretation model was the difference between bands 5 and 7. The contrast variable, computed by the Hadamard transform was significant but did not contribute much to the interpretation model. Forest areas containing very large timber volumes because of large tree sizes were not separable from areas of similar crown cover but containing smaller trees using ERTS image interpretation only. All correlations between space derived timber volume predictions and estimates obtained from aerial and ground sampling were relatively low but significant and stable. There was a much stronger relationship between variables derived from MSS and U2 data than between U2 and ground data.
Sampled-Data Techniques Applied to a Digital Controller for an Altitude Autopilot
NASA Technical Reports Server (NTRS)
Schmidt, Stanley F.; Harper, Eleanor V.
1959-01-01
Sampled-data theory, using the Z transformation, is applied to the design of a digital controller for an aircraft-altitude autopilot. Particular attention is focused on the sensitivity of the design to parameter variations and the abruptness of the response, that is, the normal acceleration required to carry out a transient maneuver. Consideration of these two characteristics of the system has shown that the finite settling time design method produces an unacceptable system, primarily because of the high sensitivity of the response to parameter variations, although abruptness can be controlled by increasing the sampling period. Also demonstrated is the importance of having well-damped poles or zeros if cancellation is attempted in the design methods. A different method of smoothing the response and obtaining a design which is not excessively sensitive is proposed, and examples are carried through to demonstrate the validity of the procedure. This method is based on design concepts of continuous systems, and it is shown that if no pole-zero cancellations are allowed in the design, one can obtain a response which is not too abrupt, is relatively insensitive to parameter variations, and is not sensitive to practical limits on control-surface rate. This particular design also has the simplest possible pulse transfer function for the digital controller. Simulation techniques and root loci are used for the verification of the design philosophy.
NASA Astrophysics Data System (ADS)
Bagchi, A.; Sarkar, S.; Mukhopadhyay, P. K.
2018-02-01
Three different coloured focused laser beams were used to study the photo induced microactuation effect found in some ferromagnetic shape memory alloys. Besides trying to uncover the basic causes of this unique and as yet unexplained effect, these studies are to help find other conditions to further characterize the effect for practical use. In this study some mechanisms have been proposed to control the amplitude of actuation of the sample. Control of the actuation of the FSMA sample both linearly with the help of a continuously variable neutral density filter as well periodically with the help of a linear polarizer was achieved. Statistical analysis of the experimental data was also done by applying ANOVA studies on the data to conclusively provide evidence in support of the relationship between the actuation of the sample and the various controlling factors. This study is expected to pave the way to implement this property of the sample in fabricating and operating useful micro-mechanical systems in the near future.
NASA Astrophysics Data System (ADS)
Bagchi, A.; Sarkar, S.; Mukhopadhyay, P. K.
2018-07-01
Three different coloured focused laser beams were used to study the photo induced microactuation effect found in some ferromagnetic shape memory alloys. Besides trying to uncover the basic causes of this unique and as yet unexplained effect, these studies are to help find other conditions to further characterize the effect for practical use. In this study some mechanisms have been proposed to control the amplitude of actuation of the sample. Control of the actuation of the FSMA sample both linearly with the help of a continuously variable neutral density filter as well periodically with the help of a linear polarizer was achieved. Statistical analysis of the experimental data was also done by applying ANOVA studies on the data to conclusively provide evidence in support of the relationship between the actuation of the sample and the various controlling factors. This study is expected to pave the way to implement this property of the sample in fabricating and operating useful micro-mechanical systems in the near future.
NASA Astrophysics Data System (ADS)
Redfern, Andrew; Koplow, Michael; Wright, Paul
2007-01-01
Most residential heating, ventilating, and air-conditioning (HVAC) systems utilize a single zone for conditioning air throughout the entire house. While inexpensive, these systems lead to wide temperature distributions and inefficient cooling due to the difference in thermal loads in different rooms. The end result is additional cost to the end user because the house is over conditioned. To reduce the total amount of energy used in a home and to increase occupant comfort there is a need for a better control system using multiple temperature zones. Typical multi-zone systems are costly and require extensive infrastructure to function. Recent advances in wireless sensor networks (WSNs) have enabled a low cost drop-in wireless vent register control system. The register control system is controlled by a master controller unit, which collects sensor data from a distributed wireless sensor network. Each sensor node samples local settings (occupancy, light, humidity and temperature) and reports the data back to the master control unit. The master control unit compiles the incoming data and then actuates the vent resisters to control the airflow throughout the house. The control system also utilizes a smart thermostat with a movable set point to enable the user to define their given comfort levels. The new system can reduce the run time of the HVAC system and thus decreasing the amount of energy used and increasing the comfort of the home occupations.
[Automatic adjustment control system for DC glow discharge plasma source].
Wan, Zhen-zhen; Wang, Yong-qing; Li, Xiao-jia; Wang, Hai-zhou; Shi, Ning
2011-03-01
There are three important parameters in the DC glow discharge process, the discharge current, discharge voltage and argon pressure in discharge source. These parameters influence each other during glow discharge process. This paper presents an automatic control system for DC glow discharge plasma source. This system collects and controls discharge voltage automatically by adjusting discharge source pressure while the discharge current is constant in the glow discharge process. The design concept, circuit principle and control program of this automatic control system are described. The accuracy is improved by this automatic control system with the method of reducing the complex operations and manual control errors. This system enhances the control accuracy of glow discharge voltage, and reduces the time to reach discharge voltage stability. The glow discharge voltage stability test results with automatic control system are provided as well, the accuracy with automatic control system is better than 1% FS which is improved from 4% FS by manual control. Time to reach discharge voltage stability has been shortened to within 30 s by automatic control from more than 90 s by manual control. Standard samples like middle-low alloy steel and tin bronze have been tested by this automatic control system. The concentration analysis precision has been significantly improved. The RSDs of all the test result are better than 3.5%. In middle-low alloy steel standard sample, the RSD range of concentration test result of Ti, Co and Mn elements is reduced from 3.0%-4.3% by manual control to 1.7%-2.4% by automatic control, and that for S and Mo is also reduced from 5.2%-5.9% to 3.3%-3.5%. In tin bronze standard sample, the RSD range of Sn, Zn and Al elements is reduced from 2.6%-4.4% to 1.0%-2.4%, and that for Si, Ni and Fe is reduced from 6.6%-13.9% to 2.6%-3.5%. The test data is also shown in this paper.
Development of a versatile user-friendly IBA experimental chamber
NASA Astrophysics Data System (ADS)
Kakuee, Omidreza; Fathollahi, Vahid; Lamehi-Rachti, Mohammad
2016-03-01
Reliable performance of the Ion Beam Analysis (IBA) techniques is based on the accurate geometry of the experimental setup, employment of the reliable nuclear data and implementation of dedicated analysis software for each of the IBA techniques. It has already been shown that geometrical imperfections lead to significant uncertainties in quantifications of IBA measurements. To minimize these uncertainties, a user-friendly experimental chamber with a heuristic sample positioning system for IBA analysis was recently developed in the Van de Graaff laboratory in Tehran. This system enhances IBA capabilities and in particular Nuclear Reaction Analysis (NRA) and Elastic Recoil Detection Analysis (ERDA) techniques. The newly developed sample manipulator provides the possibility of both controlling the tilt angle of the sample and analyzing samples with different thicknesses. Moreover, a reasonable number of samples can be loaded in the sample wheel. A comparison of the measured cross section data of the 16O(d,p1)17O reaction with the data reported in the literature confirms the performance and capability of the newly developed experimental chamber.
A computer system for analysis and transmission of spirometry waveforms using volume sampling.
Ostler, D V; Gardner, R M; Crapo, R O
1984-06-01
A microprocessor-controlled data gathering system for telemetry and analysis of spirometry waveforms was implemented using a completely digital design. Spirometry waveforms were obtained from an optical shaft encoder attached to a rolling seal spirometer. Time intervals between 10-ml volume changes (volume sampling) were stored. The digital design eliminated problems of analog signal sampling. The system measured flows up to 12 liters/sec with 5% accuracy and volumes up to 10 liters with 1% accuracy. Transmission of 10 waveforms took about 3 min. Error detection assured that no data were lost or distorted during transmission. A pulmonary physician at the central hospital reviewed the volume-time and flow-volume waveforms and interpretations generated by the central computer before forwarding the results and consulting with the rural physician. This system is suitable for use in a major hospital, rural hospital, or small clinic because of the system's simplicity and small size.
VPI - VIBRATION PATTERN IMAGER: A CONTROL AND DATA ACQUISITION SYSTEM FOR SCANNING LASER VIBROMETERS
NASA Technical Reports Server (NTRS)
Rizzi, S. A.
1994-01-01
The Vibration Pattern Imager (VPI) system was designed to control and acquire data from laser vibrometer sensors. The PC computer based system uses a digital signal processing (DSP) board and an analog I/O board to control the sensor and to process the data. The VPI system was originally developed for use with the Ometron VPI Sensor (Ometron Limited, Kelvin House, Worsley Bridge Road, London, SE26 5BX, England), but can be readily adapted to any commercially available sensor which provides an analog output signal and requires analog inputs for control of mirror positioning. VPI's graphical user interface allows the operation of the program to be controlled interactively through keyboard and mouse-selected menu options. The main menu controls all functions for setup, data acquisition, display, file operations, and exiting the program. Two types of data may be acquired with the VPI system: single point or "full field". In the single point mode, time series data is sampled by the A/D converter on the I/O board at a user-defined rate for the selected number of samples. The position of the measuring point, adjusted by mirrors in the sensor, is controlled via a mouse input. In the "full field" mode, the measurement point is moved over a user-selected rectangular area with up to 256 positions in both x and y directions. The time series data is sampled by the A/D converter on the I/O board and converted to a root-mean-square (rms) value by the DSP board. The rms "full field" velocity distribution is then uploaded for display and storage. VPI is written in C language and Texas Instruments' TMS320C30 assembly language for IBM PC series and compatible computers running MS-DOS. The program requires 640K of RAM for execution, and a hard disk with 10Mb or more of disk space is recommended. The program also requires a mouse, a VGA graphics display, a Four Channel analog I/O board (Spectrum Signal Processing, Inc.; Westborough, MA), a break-out box and a Spirit-30 board (Sonitech International, Inc.; Wellesley, MA) which includes a TMS320C30 DSP processor, 256Kb zero wait state SRAM, and a daughter board with 8Mb one wait state DRAM. Please contact COSMIC for additional information on required hardware and software. In order to compile the provided VPI source code, a Microsoft C version 6.0 compiler, a Texas Instruments' TMS320C30 assembly language compiler, and the Spirit 30 run time libraries are required. A math co-processor is highly recommended. A sample MS-DOS executable is provided on the distribution medium. The standard distribution medium for this program is one 5.25 inch 360K MS-DOS format diskette. The contents of the diskettes are compressed using the PKWARE archiving tools. The utility to unarchive the files, PKUNZIP.EXE, is included. VPI was developed in 1991-1992.
NASA Astrophysics Data System (ADS)
Yin, Bo; Liu, Li; Wang, Jiahan; Li, Xiran; Liu, Zhenbo; Li, Dewei; Wang, Jun; Liu, Lu; Wu, Jun; Xu, Tingting; Cui, He
2017-10-01
Electric energy measurement as a basic work, an accurate measurements play a vital role for the economic interests of both parties of power supply, the standardized management of the measurement laboratory at all levels is a direct factor that directly affects the fairness of measurement. Currently, the management of metering laboratories generally uses one-dimensional bar code as the recognition object, advances the testing process by manual management, most of the test data requires human input to generate reports. There are many problems and potential risks in this process: Data cannot be saved completely, cannot trace the status of inspection, the inspection process isn't completely controllable and so on. For the provincial metrology center's actual requirements of the whole process management for the performance test of the power measuring appliances, using of large-capacity RF tags as a process management information media, we developed a set of general measurement experiment management system, formulated a standardized full performance test process, improved the raw data recording mode of experimental process, developed a storehouse automatic inventory device, established a strict test sample transfer and storage system, ensured that all the raw data of the inspection can be traced back, achieved full life-cycle control of the sample, significantly improved the quality control level and the effectiveness of inspection work.
Automated Microflow NMR: Routine Analysis of Five-Microliter Samples
Jansma, Ariane; Chuan, Tiffany; Geierstanger, Bernhard H.; Albrecht, Robert W.; Olson, Dean L.; Peck, Timothy L.
2006-01-01
A microflow CapNMR probe double-tuned for 1H and 13C was installed on a 400-MHz NMR spectrometer and interfaced to an automated liquid handler. Individual samples dissolved in DMSO-d6 are submitted for NMR analysis in vials containing as little as 10 μL of sample. Sets of samples are submitted in a low-volume 384-well plate. Of the 10 μL of sample per well, as with vials, 5 μL is injected into the microflow NMR probe for analysis. For quality control of chemical libraries, 1D NMR spectra are acquired under full automation from 384-well plates on as many as 130 compounds within 24 h using 128 scans per spectrum and a sample-to-sample cycle time of ∼11 min. Because of the low volume requirements and high mass sensitivity of the microflow NMR system, 30 nmol of a typical small molecule is sufficient to obtain high-quality, well-resolved, 1D proton or 2D COSY NMR spectra in ∼6 or 20 min of data acquisition time per experiment, respectively. Implementation of pulse programs with automated solvent peak identification and suppression allow for reliable data collection, even for samples submitted in fully protonated DMSO. The automated microflow NMR system is controlled and monitored using web-based software. PMID:16194121
Snowpack Chemistry of Reactive Gases at Station Concordia, Antarctica
NASA Astrophysics Data System (ADS)
Helmig, Detlev; Mass, Alex; Hueber, Jacques; Fain, Xavier; Dommergue, Aurelien; Barbero, Albane; Savarino, Joel
2013-04-01
During December 2012 a new experiment for the study of snow photochemical processes and surface gas exchange was installed at Dome Concordia, Antarctica. The experiment consists of two sampling manifolds ('snow tower') which facilitate the withdrawal of interstitial firn air from four depths in the snowpack and from above the surface. One of these snow towers can be shaded for investigation of the dependency of snow chemistry on solar radiation. A nearby 12 m meteorological tower facilitates above surface turbulence and trace gas gradient measurements. Temperature profiles and UV and IR light penetration are monitored in the snowpack. Air samples are directed through sampling lines to a nearby underground laboratory that houses the experiment control system and gas monitors. The system is fully automated, sampling gases from the array of inlet ports sequentially, and is intended to be operated continuously for a full annual cycle. The computerized control system can be accessed remotely for data retrieval and quality control and for configuring experimental details. Continuous gas measurements include ozone, nitrogen oxides, methane, carbon monoxide, and gaseous elemental mercury. Whole air samples were sampled on four occasions for volatile organic compound analysis. The objective of this research is the study of the year-round snowpack gas chemistry and its dependency on snowpack and above surface physical and environmental conditions. A particular emphasis will be the investigation of the effects of increased UV radiation during the occurrence of the stratospheric ozone hole. We will present the conceptual design of the experiment and data examples from the first three months of the experiment.
Time-Domain Terahertz Computed Axial Tomography NDE System
NASA Technical Reports Server (NTRS)
Zimdars, David
2012-01-01
NASA has identified the need for advanced non-destructive evaluation (NDE) methods to characterize aging and durability in aircraft materials to improve the safety of the nation's airline fleet. 3D THz tomography can play a major role in detection and characterization of flaws and degradation in aircraft materials, including Kevlar-based composites and Kevlar and Zylon fabric covers for soft-shell fan containment where aging and durability issues are critical. A prototype computed tomography (CT) time-domain (TD) THz imaging system has been used to generate 3D images of several test objects including a TUFI tile (a thermal protection system tile used on the Space Shuttle and possibly the Orion or similar capsules). This TUFI tile had simulated impact damage that was located and the depth of damage determined. The CT motion control gan try was designed and constructed, and then integrated with a T-Ray 4000 control unit and motion controller to create a complete CT TD-THz imaging system prototype. A data collection software script was developed that takes multiple z-axis slices in sequence and saves the data for batch processing. The data collection software was integrated with the ability to batch process the slice data with the CT TD-THz image reconstruction software. The time required to take a single CT slice was decreased from six minutes to approximately one minute by replacing the 320 ps, 100-Hz waveform acquisition system with an 80 ps, 1,000-Hz waveform acquisition system. The TD-THZ computed tomography system was built from pre-existing commercial off-the-shelf subsystems. A CT motion control gantry was constructed from COTS components that can handle larger samples. The motion control gantry allows inspection of sample sizes of up to approximately one cubic foot (.0.03 cubic meters). The system reduced to practice a CT-TDTHz system incorporating a COTS 80- ps/l-kHz waveform scanner. The incorporation of this scanner in the system allows acquisition of 3D slice data with better signal-to-noise using a COTS scanner rather than the gchirped h scanner. The system also reduced to practice a prototype for commercial CT systems for insulating materials where safety concerns cannot accommodate x-ray. A software script was written to automate the COTS software to collect and process TD-THz CT data.
1985-09-01
Transducers capable of measuring electro-hydraulic control system which fore-aft and vertical load on a driven controls the brake system to deactivate tire...power. * axle allows design of all load-carrying - System logic power. ENGINE I EXTERNAL COMPARTMENT COMPONENTS CAB Brake Levelin system I trans... brake con- The TWS DAS was designed to 1) pro- trol system . vide onboard data sampling and filtering, A simplified truck operational flow chart 2) make
NASA Technical Reports Server (NTRS)
SanSoucie, Michael P.; Vermilion, David J.; Rogers, Jan R.
2015-01-01
The NASA Marshall Space Flight Center (MSFC) electrostatic levitation (ESL) laboratory has a long history of providing materials research and thermophysical property data. A summary of the labs capabilities, recent upgrades, and ongoing and future work will be provided. The laboratory has recently added two new capabilities to its main levitation chamber: a rapid quench system and an oxygen control system. The rapid quench system allows samples to be dropped into a quench vessel that can be filled with a low melting point material, such as a gallium or indium alloy. Thereby allowing rapid quenching of undercooled liquid metals. The oxygen control system consists of an oxygen sensor, oxygen pump, and a control unit. The sensor is a potentiometric device that determines the difference in oxygen activity between two gas compartments separated by an electrolyte, which is yttria-stabilized zirconia. The pump utilizes coulometric titration to either add or remove oxygen. The system is controlled by a desktop control unit, which can also be accessed via a computer. This system allows the oxygen partial pressure within the vacuum chamber to be measured and controlled, theoretically in the range from 10-36 to 100 bar. The ESL laboratory also has an emissometer, called the High-Temperature Emissivity Measurement System (HiTEMS). This system measures the spectral emissivity of materials from 600degC to 3,000degC. The system consists of a vacuum chamber, a black body source, and a Fourier Transform Infrared Spectrometer (FTIR). The system utilizes optics to swap the signal between the sample and the black body. The system was originally designed to measure the hemispherical spectral emissivity of levitated samples, which are typically 2.5mm spheres. Levitation allows emissivity measurements of molten samples, but more work is required to develop this capability. The system is currently setup measure the near-normal spectral emissivity of stationary samples, which has been used to take measurements of ablative materials, rocket nozzle coating materials, and materials for spacecraft instruments.
NASA Astrophysics Data System (ADS)
Beddow, B.; Roberts, C.; Rankin, J.; Bloch, A.; Peizer, J.
1981-01-01
The National Accident Sampling System (NASS) is described. The study area discussed is one of the original ten sites selected for NASS implementation. In addition to collecting data from the field, the original ten sites address questions of feasibility of the plan, projected results of the data collection effort, and specific operational topics, e.g., team size, sampling requirements, training approaches, quality control procedures, and field techniques. Activities and results of the first three years of the project, for both major tasks (establishment and operation) are addressed. Topics include: study area documentation; team description, function and activities; problems and solutions; and recommendations.
Automated Test Systems for Toxic Vapor Detectors
NASA Technical Reports Server (NTRS)
Mattson, C. B.; Hammond, T. A.; Schwindt, C. J.
1997-01-01
The NASA Toxic Vapor Detection Laboratory (TVDL) at the Kennedy Space Center (KSC), Florida, has been using Personal Computer based Data Acquisition and Control Systems (PCDAS) for about nine years. These systems control the generation of toxic vapors of known concentrations under controlled conditions of temperature and humidity. The PCDAS also logs the test conditions and the test article responses in data files for analysis by standard spreadsheets or custom programs. The PCDAS was originally developed to perform standardized qualification and acceptance tests in a search for a commercial off-the-shelf (COTS) toxic vapor detector to replace the hydrazine detectors for the Space Shuttle launch pad. It has since become standard test equipment for the TVDL and is indispensable in producing calibration standards for the new hydrazine monitors at the 10 part per billion (ppb) level. The standard TVDL PCDAS can control two toxic vapor generators (TVG's) with three channels each and two flow/ temperature / humidity (FTH) controllers and it can record data from up to six toxic vapor detectors (TVD's) under test and can deliver flows from 5 to 50 liters per minute (L/m) at temperatures from near zero to 50 degrees Celsius (C) using an environmental chamber to maintain the sample temperature. The concentration range for toxic vapors depends on the permeation source installed in the TVG. The PCDAS can provide closed loop control of temperature and humidity to two sample vessels, typically one for zero gas and one for the standard gas. This is required at very low toxic vapor concentrations to minimize the time required to passivate the sample delivery system. Recently, there have been several requests for information about the PCDAS by other laboratories with similar needs, both on and off KSC. The purpose of this paper is to inform the toxic vapor detection community of the current status and planned upgrades to the automated testing of toxic vapor detectors at the Kennedy Space Center.
Automated Test Systems for Toxic Vapor Detectors
NASA Technical Reports Server (NTRS)
Mattson, C. B.; Hammond, T. A.; Schwindt, C. J.
1997-01-01
The NASA Toxic Vapor Detection Laboratory (TVDL) at the Kennedy Space Center (KSC), Florida, has been using Personal Computer based Data Acquisition and Control Systems (PCDAS) for about nine years. These systems control the generation of toxic vapors of known concentrations under controlled conditions of temperature and humidity. The PCDAS also logs the test conditions and the test article responses in data files for analysis by standard spreadsheets or custom programs. The PCDAS was originally developed to perform standardized qualification and acceptance tests in a search for a commercial off-the-shelf (COTS) toxic vapor detector to replace the hydrazine detectors for the Space Shuttle launch pad. It has since become standard test equipment for the TVDL and is indispensable in producing calibration standards for the new hydrazine monitors at the 10 part per billion (ppb) level. The standard TVDL PCDAS can control two toxic vapor generators (TVG's) with three channels each and two flow/temperature/humidity (FIFH) controllers and it can record data from up to six toxic vapor detectors (TVD's) under test and can deliver flows from 5 to 50 liters per minute (L/m) at temperatures from near zero to 50 degrees Celsius (C) using an environmental chamber to maintain the sample temperature. The concentration range for toxic vapors depends on the permeation source installed in the TVG. The PCDAS can provide closed loop control of temperature and humidity to two sample vessels, typically one for zero gas and one for the standard gas. This is required at very low toxic vapor concentrations to minimize the time required to passivate the sample delivery system. Recently, there have been several requests for information about the PCDAS by other laboratories with similar needs, both on and off KSC. The purpose of this paper is to inform the toxic vapor detection community of the current status and planned upgrades to the automated testing of toxic vapor detectors at the Kennedy Space Center.
Self-tuning control of attitude and momentum management for the Space Station
NASA Technical Reports Server (NTRS)
Shieh, L. S.; Sunkel, J. W.; Yuan, Z. Z.; Zhao, X. M.
1992-01-01
This paper presents a hybrid state-space self-tuning design methodology using dual-rate sampling for suboptimal digital adaptive control of attitude and momentum management for the Space Station. This new hybrid adaptive control scheme combines an on-line recursive estimation algorithm for indirectly identifying the parameters of a continuous-time system from the available fast-rate sampled data of the inputs and states and a controller synthesis algorithm for indirectly finding the slow-rate suboptimal digital controller from the designed optimal analog controller. The proposed method enables the development of digitally implementable control algorithms for the robust control of Space Station Freedom with unknown environmental disturbances and slowly time-varying dynamics.
Chandler, Mark A.; Goggin, David J.; Horne, Patrick J.; Kocurek, Gary G.; Lake, Larry W.
1989-01-01
For making rapid, non-destructive permeability measurements in the field, a portable minipermeameter of the kind having a manually-operated gas injection tip is provided with a microcomputer system which operates a flow controller to precisely regulate gas flow rate to a test sample, and reads a pressure sensor which senses the pressure across the test sample. The microcomputer system automatically turns on the gas supply at the start of each measurement, senses when a steady-state is reached, collects and records pressure and flow rate data, and shuts off the gas supply immediately after the measurement is completed. Preferably temperature is also sensed to correct for changes in gas viscosity. The microcomputer system may also provide automatic zero-point adjustment, sensor calibration, over-range sensing, and may select controllers, sensors, and set-points for obtaining the most precise measurements. Electronic sensors may provide increased accuracy and precision. Preferably one microcomputer is used for sensing instrument control and data collection, and a second microcomputer is used which is dedicated to recording and processing the data, selecting the sensors and set-points for obtaining the most precise measurements, and instructing the user how to set-up and operate the minipermeameter. To provide mass data collection and user-friendly operation, the second microcomputer is preferably a lap-type portable microcomputer having a non-volatile or battery-backed CMOS memory.
Skinner, Kenneth D.
2018-05-11
Groundwater-quality samples and water-level data were collected from 36 wells in the Jerome/Gooding County area of the eastern Snake River Plain aquifer during June 2017. The wells included 30 wells sampled for the U.S. Geological Survey’s National Water-Quality Assessment project, plus an additional 6 wells were selected to increase spatial distribution. The data provide water managers with the ability for an improved understanding of groundwater quality and flow directions in the area. Groundwater-quality samples were analyzed for nutrients, major ions, trace elements, and stable isotopes of water. Quality-assurance and quality-control measures consisted of multiple blank samples and a sequential replicate sample. All data are available online at the USGS National Water Information System.
Evaluation of measurement data from a sensor system for breath control
NASA Astrophysics Data System (ADS)
Seifert, Rolf; Keller, Hubert B.; Conrad, Thorsten; Peter, Jens
2017-03-01
Binary ethanol-H2 gas samples were measured by an innovative mobile sensor system for the alcohol control in the respiratory air. The measurements were performed by a gas sensor operated by cyclic variation of the working temperature at the sensor head. The evaluation of the data, using an updated version of the evaluation procedure ProSens, results in a very good substance identification and concentration determination of the components of the gas mixture. The relative analysis errors were in all cases less than 9%.
An Indoor Location-Based Control System Using Bluetooth Beacons for IoT Systems.
Huh, Jun-Ho; Seo, Kyungryong
2017-12-19
The indoor location-based control system estimates the indoor position of a user to provide the service he/she requires. The major elements involved in the system are the localization server, service-provision client, user application positioning technology. The localization server controls access of terminal devices (e.g., Smart Phones and other wireless devices) to determine their locations within a specified space first and then the service-provision client initiates required services such as indoor navigation and monitoring/surveillance. The user application provides necessary data to let the server to localize the devices or allow the user to receive various services from the client. The major technological elements involved in this system are indoor space partition method, Bluetooth 4.0, RSSI (Received Signal Strength Indication) and trilateration. The system also employs the BLE communication technology when determining the position of the user in an indoor space. The position information obtained is then used to control a specific device(s). These technologies are fundamental in achieving a "Smart Living". An indoor location-based control system that provides services by estimating user's indoor locations has been implemented in this study (First scenario). The algorithm introduced in this study (Second scenario) is effective in extracting valid samples from the RSSI dataset but has it has some drawbacks as well. Although we used a range-average algorithm that measures the shortest distance, there are some limitations because the measurement results depend on the sample size and the sample efficiency depends on sampling speeds and environmental changes. However, the Bluetooth system can be implemented at a relatively low cost so that once the problem of precision is solved, it can be applied to various fields.
An Indoor Location-Based Control System Using Bluetooth Beacons for IoT Systems
Huh, Jun-Ho; Seo, Kyungryong
2017-01-01
The indoor location-based control system estimates the indoor position of a user to provide the service he/she requires. The major elements involved in the system are the localization server, service-provision client, user application positioning technology. The localization server controls access of terminal devices (e.g., Smart Phones and other wireless devices) to determine their locations within a specified space first and then the service-provision client initiates required services such as indoor navigation and monitoring/surveillance. The user application provides necessary data to let the server to localize the devices or allow the user to receive various services from the client. The major technological elements involved in this system are indoor space partition method, Bluetooth 4.0, RSSI (Received Signal Strength Indication) and trilateration. The system also employs the BLE communication technology when determining the position of the user in an indoor space. The position information obtained is then used to control a specific device(s). These technologies are fundamental in achieving a “Smart Living”. An indoor location-based control system that provides services by estimating user’s indoor locations has been implemented in this study (First scenario). The algorithm introduced in this study (Second scenario) is effective in extracting valid samples from the RSSI dataset but has it has some drawbacks as well. Although we used a range-average algorithm that measures the shortest distance, there are some limitations because the measurement results depend on the sample size and the sample efficiency depends on sampling speeds and environmental changes. However, the Bluetooth system can be implemented at a relatively low cost so that once the problem of precision is solved, it can be applied to various fields. PMID:29257044
NASA Astrophysics Data System (ADS)
Leifer, R.; Sommers, K. G.; Guggenheim, S. F.; Fisenne, I.
1981-02-01
An ultra-clean, low volume gas sampling system (CLASS), flown aboard a high altitude aircraft (WB-57F), and providing information on stratospheric trace gases is presented. Attention is given to the instrument design and the electronic control design. Since remote operation is mandatory on the WB-57F, a servo pressure transducer, electrical pressure switch for automatic shutdown, and a mechanical safety relief valve were installed on the sampling manifold, indicated on the CLASS flow chart. The electronic control system consists of hermetically sealed solid state timers, relays, and a stepping switch, for controlling the compressor pump and solenoid valves. In designing the automatic control system, vibration, shock, acceleration, extreme low temperature, and aircraft safety were important considerations. CLASS was tested on three separate occasions, and tables of analytical data from these flights are presented. Readiness capability was demonstrated when the Mount St. Helens eruption plume of May 18, 1980, was intercepted, and it was concluded that no large injection of Rn-222 entered the stratosphere or troposphere from the eruption.
Interoperability of Information Systems Managed and Used by the Local Health Departments.
Shah, Gulzar H; Leider, Jonathon P; Luo, Huabin; Kaur, Ravneet
2016-01-01
In the post-Affordable Care Act era marked by interorganizational collaborations and availability of large amounts of electronic data from other community partners, it is imperative to assess the interoperability of information systems used by the local health departments (LHDs). To describe the level of interoperability of LHD information systems and identify factors associated with lack of interoperability. This mixed-methods research uses data from the 2015 Informatics Capacity and Needs Assessment Survey, with a target population of all LHDs in the United States. A representative sample of 650 LHDs was drawn using a stratified random sampling design. A total of 324 completed responses were received (50% response rate). Qualitative data were used from a key informant interview study of LHD informatics staff from across the United States. Qualitative data were independently coded by 2 researchers and analyzed thematically. Survey data were cleaned, bivariate comparisons were conducted, and a multivariable logistic regression was run to characterize factors associated with interoperability. For 30% of LHDs, no systems were interoperable, and 38% of LHD respondents indicated some of the systems were interoperable. Significant determinants of interoperability included LHDs having leadership support (adjusted odds ratio [AOR] = 3.54), control of information technology budget allocation (AOR = 2.48), control of data systems (AOR = 2.31), having a strategic plan for information systems (AOR = 1.92), and existence of business process analysis and redesign (AOR = 1.49). Interoperability of all systems may be an informatics goal, but only a small proportion of LHDs reported having interoperable systems, pointing to a substantial need among LHDs nationwide.
Xiong, Wenjun; Patel, Ragini; Cao, Jinde; Zheng, Wei Xing
In this brief, our purpose is to apply asynchronous and intermittent sampled-data control methods to achieve the synchronization of hierarchical time-varying neural networks. The asynchronous and intermittent sampled-data controllers are proposed for two reasons: 1) the controllers may not transmit the control information simultaneously and 2) the controllers cannot always exist at any time . The synchronization is then discussed for a kind of hierarchical time-varying neural networks based on the asynchronous and intermittent sampled-data controllers. Finally, the simulation results are given to illustrate the usefulness of the developed criteria.In this brief, our purpose is to apply asynchronous and intermittent sampled-data control methods to achieve the synchronization of hierarchical time-varying neural networks. The asynchronous and intermittent sampled-data controllers are proposed for two reasons: 1) the controllers may not transmit the control information simultaneously and 2) the controllers cannot always exist at any time . The synchronization is then discussed for a kind of hierarchical time-varying neural networks based on the asynchronous and intermittent sampled-data controllers. Finally, the simulation results are given to illustrate the usefulness of the developed criteria.
New controlled environment vitrification system for preparing wet samples for cryo-SEM.
Ge, H; Suszynski, W J; Davis, H T; Scriven, L E
2008-01-01
A new controlled environment vitrification system (CEVS) has been designed and constructed to facilitate examination by cryogenic scanning electron microscopy (Cryo-SEM) of initial suspension state and of microstructure development in latex, latex-composite and other coatings while they still contain solvent. The new system has a main chamber with provisions for coating as well as drying, and for well-controlled plunging into cryogen. An added subsidiary chamber holds samples for drying or annealing over minutes to days before they are returned to the main chamber and plunged from it. In the main chamber, samples are blade-coated on 5 x 7 mm pieces of silicon wafer and held at selected temperature and humidity for successively longer times, either there or after transfer along a rail into the subsidiary chamber. They are then placed in the sample holder mounted on the plunge rod, so as to permit adjustment of the sample's attitude when it plunges, at controlled speed, into liquid ethane at its freezing point, to a chosen depth, in order to solidify the sample without significant shear or freezing artifacts. The entries of plunging samples and related sample holders into liquid ethane were recorded with a high-speed, high-resolution Photron digital camera. The data were interpreted with a new hypothesis about the width of the band of extremely rapid cooling by deeply subcooled nucleate boiling below the line of entry. Complementary cryo-SEM images revealed that the freezing rate and surface shearing of a sample need to be balanced by adjusting the plunging attitude.
Integrating laboratory robots with analytical instruments--must it really be so difficult?
Kramer, G W
1990-09-01
Creating a reliable system from discrete laboratory instruments is often a task fraught with difficulties. While many modern analytical instruments are marvels of detection and data handling, attempts to create automated analytical systems incorporating such instruments are often frustrated by their human-oriented control structures and their egocentricity. The laboratory robot, while fully susceptible to these problems, extends such compatibility issues to the physical dimensions involving sample interchange, manipulation, and event timing. The workcell concept was conceived to describe the procedure and equipment necessary to carry out a single task during sample preparation. This notion can be extended to organize all operations in an automated system. Each workcell, no matter how complex its local repertoire of functions, must be minimally capable of accepting information (commands, data), returning information on demand (status, results), and being started, stopped, and reset by a higher level device. Even the system controller should have a mode where it can be directed by instructions from a higher level.
Pump efficiency in solar-energy systems
NASA Technical Reports Server (NTRS)
1978-01-01
Study investigates characteristics of typical off-the-shelf pumping systems that might be used in solar systems. Report includes discussion of difficulties in predicting pump efficiency from manufacturers' data. Sample calculations are given. Peak efficiencies, flow-rate control, and noise levels are investigated. Review or theory of pumps types and operating characteristics is presented.
Automation--down to the nuts and bolts.
Fix, R J; Rowe, J M; McConnell, B C
2000-01-01
Laboratories that once viewed automation as an expensive luxury are now looking to automation as a solution to increase sample throughput, to help ensure data integrity and to improve laboratory safety. The question is no longer, 'Should we automate?', but 'How should we approach automation?' A laboratory may choose from three approaches when deciding to automate: (1) contract with a third party vendor to produce a turnkey system, (2) develop and fabricate the system in-house or (3) some combination of approaches (1) and (2). The best approach for a given laboratory depends upon its available resources. The first lesson to be learned in automation is that no matter how straightforward an idea appears in the beginning, the solution will not be realized until many complex problems have been resolved. Issues dealing with sample vessel manipulation, liquid handling and system control must be addressed before a final design can be developed. This requires expertise in engineering, electronics, programming and chemistry. Therefore, the team concept of automation should be employed to help ensure success. This presentation discusses the advantages and disadvantages of the three approaches to automation. The development of an automated sample handling and control system for the STAR System focused microwave will be used to illustrate the complexities encountered in a seemingly simple project, and to highlight the importance of the team concept to automation no matter which approach is taken. The STAR System focused microwave from CEM Corporation is an open vessel digestion system with six microwave cells. This system is used to prepare samples for trace metal determination. The automated sample handling was developed around a XYZ motorized gantry system. Grippers were specially designed to perform several different functions and to provide feedback to the control software. Software was written in Visual Basic 5.0 to control the movement of the samples and the operation and monitoring of the STAR microwave. This software also provides a continuous update of the system's status to the computer screen. The system provides unattended preparation of up to 59 samples per run.
NASA Astrophysics Data System (ADS)
Manzella, G. M. R.; Scoccimarro, E.; Pinardi, N.; Tonani, M.
2003-01-01
A "ship of opportunity" program was launched as part of the Mediterranean Forecasting System Pilot Project. During the operational period (September 1999 to May 2000), six tracks covered the Mediterranean from the northern to southern boundaries approximately every 15 days, while a long eastwest track from Haifa to Gibraltar was covered approximately every month. XBT data were collected, sub-sampled at 15 inflection points and transmitted through a satellite communication system to a regional data centre. It was found that this data transmission system has limitations in terms of quality of the temperature profiles and quantity of data successfully transmitted. At the end of the MFSPP operational period, a new strategy for data transmission and management was developed. First of all, VOS-XBT data are transmitted with full resolution. Secondly, a new data management system, called Near Real Time Quality Control for XBT (NRT.QC.XBT), was defined to produce a parallel stream of high quality XBT data for further scientific analysis. The procedure includes: (1) Position control; (2) Elimination of spikes; (3) Re-sampling at a 1 metre vertical interval; (4) Filtering; (5) General malfunctioning check; (6) Comparison with climatology (and distance from this in terms of standard deviations); (7) Visual check; and (8) Data consistency check. The first six steps of the new procedure are completely automated; they are also performed using a new climatology developed as part of the project. The visual checks are finally done with a free-market software that allows NRT final data assessment.
NASA Technical Reports Server (NTRS)
Taylor, R. B.; Zwicke, P. E.; Gold, P.; Miao, W.
1980-01-01
An analytical study was conducted to define the basic configuration of an active control system for helicopter vibration and gust response alleviation. The study culminated in a control system design which has two separate systems: narrow band loop for vibration reduction and wider band loop for gust response alleviation. The narrow band vibration loop utilizes the standard swashplate control configuration to input controller for the vibration loop is based on adaptive optimal control theory and is designed to adapt to any flight condition including maneuvers and transients. The prime characteristics of the vibration control system is its real time capability. The gust alleviation control system studied consists of optimal sampled data feedback gains together with an optimal one-step-ahead prediction. The prediction permits the estimation of the gust disturbance which can then be used to minimize the gust effects on the helicopter.
Hamilton, Kristen R.; Sinha, Rajita; Potenza, Marc N.
2014-01-01
Stress has been associated with poor self-control. Individual differences in impulsivity and other behavioral tendencies may influence the relationship of stress with self-control, although this possibility has not been examined to date. The present research investigated whether cumulative stress is associated with poor self-control, and whether this relationship is mediated by impulsivity, behavioral approach, and behavioral inhibition in men and women. A community sample of 566 adults (319 women and 247 men) was assessed on the Cumulative Adversity Interview, Brief Self-control Scale, Barratt Impulsivity Scale, and Behavioral Activation System and Behavioral Inhibition System Scale (BIS/BAS). Data were analyzed using regression and bootstrapping techniques. In the total sample, the effects of cumulative stress on self-control were mediated by impulsivity. Neither behavioral inhibition nor behavioral approach mediated the association between cumulative stress and self-control in the total sample. Results were similar when men and women were considered separately, with impulsivity, but not behavioral inhibition or approach, mediating the association between cumulative stress and self-control. Impulsive individuals might benefit preferentially from interventions focusing on stress management and strategies for improving self-control. PMID:24508183
Hamilton, Kristen R; Sinha, Rajita; Potenza, Marc N
2014-11-01
Stress has been associated with poor self-control. Individual differences in impulsivity and other behavioral tendencies may influence the relationship of stress with self-control, although this possibility has not been examined to date. The present research investigated whether cumulative stress is associated with poor self-control, and whether this relationship is mediated by impulsivity, behavioral approach, and behavioral inhibition in men and women. A community sample of 566 adults (319 women and 247 men) was assessed on the Cumulative Adversity Interview, Brief Self-control Scale, Barratt Impulsivity Scale, and Behavioral Activation System and Behavioral Inhibition System Scale (BIS/BAS). Data were analyzed using regression and bootstrapping techniques. In the total sample, the effects of cumulative stress on self-control were mediated by impulsivity. Neither behavioral inhibition nor behavioral approach mediated the association between cumulative stress and self-control in the total sample. Results were similar when men and women were considered separately, with impulsivity, but not behavioral inhibition or approach, mediating the association between cumulative stress and self-control. Impulsive individuals might benefit preferentially from interventions focusing on stress management and strategies for improving self-control. Copyright © 2014 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Sui, Xin; Yang, Yongqing; Xu, Xianyun; Zhang, Shuai; Zhang, Lingzhong
2018-02-01
This paper investigates the consensus of multi-agent systems with probabilistic time-varying delays and packet losses via sampled-data control. On the one hand, a Bernoulli-distributed white sequence is employed to model random packet losses among agents. On the other hand, a switched system is used to describe packet dropouts in a deterministic way. Based on the special property of the Laplacian matrix, the consensus problem can be converted into a stabilization problem of a switched system with lower dimensions. Some mean square consensus criteria are derived in terms of constructing an appropriate Lyapunov function and using linear matrix inequalities (LMIs). Finally, two numerical examples are given to show the effectiveness of the proposed method.
The 1985 Army Experience Survey: Tabular Descriptions of First-Term Attritees. Volume 1
1986-01-01
Assistance. Survey data were processed through survey receipt control and sample management systems . Data were also keyed, edited, coded, and weighted. The...270-271 048 R136 REGION OF RESIDENCE WHEN YOU JOINED ARMY ...... .................. ... 272-273 049 E137 # TERMS OF... system - Weed out poor soldiers - Could not get promoted - Poor Day/bonuses/benefits - Other 15: Patriotism & defense - Patriotic fulfillment - Service
Autonomous GN and C for Spacecraft Exploration of Comets and Asteroids
NASA Technical Reports Server (NTRS)
Carson, John M.; Mastrodemos, Nickolaos; Myers, David M.; Acikmese, Behcet; Blackmore, James C.; Moussalis, Dhemetrio; Riedel, Joseph E.; Nolet, Simon; Chang, Johnny T.; Mandic, Milan;
2010-01-01
A spacecraft guidance, navigation, and control (GN&C) system is needed to enable a spacecraft to descend to a surface, take a sample using a touch-and-go (TAG) sampling approach, and then safely ascend. At the time of this reporting, a flyable GN&C system that can accomplish these goals is beyond state of the art. This article describes AutoGNC, which is a GN&C system capable of addressing these goals, which has recently been developed and demonstrated to a maturity TRL-5-plus. The AutoGNC solution matures and integrates two previously existing JPL capabilities into a single unified GN&C system. The two capabilities are AutoNAV and GREX. AutoNAV is JPL s current flight navigation system, and is fairly mature with respect to flybys and rendezvous with small bodies, but is lacking capability for close surface proximity operations, sampling, and contact. G-REX is a suite of low-TRL algorithms and capabilities that enables spacecraft operations in close surface proximity and for performing sampling/contact. The development and integration of AutoNAV and G-REX components into AutoGNC provides a single, unified GN&C capability for addressing the autonomy, close-proximity, and sampling/contact aspects of small-body sample return missions. AutoGNC is an integrated capability comprising elements that were developed separately. The main algorithms and component capabilities that have been matured and integrated are autonomy for near-surface operations, terrain-relative navigation (TRN), real-time image-based feedback guidance and control, and six degrees of freedom (6DOF) control of the TAG sampling event. Autonomy is achieved based on an AutoGNC Executive written in Virtual Machine Language (VML) incorporating high-level control, data management, and fault protection. In descending to the surface, the AutoGNC system uses camera images to determine its position and velocity relative to the terrain. This capability for TRN leverages native capabilities of the original AutoNAV system, but required advancements that integrate the separate capabilities for shape modeling, state estimation, image rendering, defining a database of onboard maps, and performing real-time landmark recognition against the stored maps. The ability to use images to guide the spacecraft requires the capability for image-based feedback control. In Auto- GNC, navigation estimates are fed into an onboard guidance and control system that keeps the spacecraft guided along a desired path, as it descends towards its targeted landing or sampling site. Once near the site, AutoGNC achieves a prescribed guidance condition for TAG sampling (position/orientation, velocity), and a prescribed force profile on the sampling end-effector. A dedicated 6DOF TAG control then implements the ascent burn while recovering from sampling disturbances and induced attitude rates. The control also minimizes structural interactions with flexible solar panels and disallows any part of the spacecraft from making contact with the ground (other than the intended end-effector).
Interoperability of Information Systems Managed and Used by the Local Health Departments
Leider, Jonathon P.; Luo, Huabin; Kaur, Ravneet
2016-01-01
Background: In the post-Affordable Care Act era marked by interorganizational collaborations and availability of large amounts of electronic data from other community partners, it is imperative to assess the interoperability of information systems used by the local health departments (LHDs). Objectives: To describe the level of interoperability of LHD information systems and identify factors associated with lack of interoperability. Data and Methods: This mixed-methods research uses data from the 2015 Informatics Capacity and Needs Assessment Survey, with a target population of all LHDs in the United States. A representative sample of 650 LHDs was drawn using a stratified random sampling design. A total of 324 completed responses were received (50% response rate). Qualitative data were used from a key informant interview study of LHD informatics staff from across the United States. Qualitative data were independently coded by 2 researchers and analyzed thematically. Survey data were cleaned, bivariate comparisons were conducted, and a multivariable logistic regression was run to characterize factors associated with interoperability. Results: For 30% of LHDs, no systems were interoperable, and 38% of LHD respondents indicated some of the systems were interoperable. Significant determinants of interoperability included LHDs having leadership support (adjusted odds ratio [AOR] = 3.54), control of information technology budget allocation (AOR = 2.48), control of data systems (AOR = 2.31), having a strategic plan for information systems (AOR = 1.92), and existence of business process analysis and redesign (AOR = 1.49). Conclusion: Interoperability of all systems may be an informatics goal, but only a small proportion of LHDs reported having interoperable systems, pointing to a substantial need among LHDs nationwide. PMID:27684616
Fault Isolation Filter for Networked Control System with Event-Triggered Sampling Scheme
Li, Shanbin; Sauter, Dominique; Xu, Bugong
2011-01-01
In this paper, the sensor data is transmitted only when the absolute value of difference between the current sensor value and the previously transmitted one is greater than the given threshold value. Based on this send-on-delta scheme which is one of the event-triggered sampling strategies, a modified fault isolation filter for a discrete-time networked control system with multiple faults is then implemented by a particular form of the Kalman filter. The proposed fault isolation filter improves the resource utilization with graceful fault estimation performance degradation. An illustrative example is given to show the efficiency of the proposed method. PMID:22346590
Data-driven process decomposition and robust online distributed modelling for large-scale processes
NASA Astrophysics Data System (ADS)
Shu, Zhang; Lijuan, Li; Lijuan, Yao; Shipin, Yang; Tao, Zou
2018-02-01
With the increasing attention of networked control, system decomposition and distributed models show significant importance in the implementation of model-based control strategy. In this paper, a data-driven system decomposition and online distributed subsystem modelling algorithm was proposed for large-scale chemical processes. The key controlled variables are first partitioned by affinity propagation clustering algorithm into several clusters. Each cluster can be regarded as a subsystem. Then the inputs of each subsystem are selected by offline canonical correlation analysis between all process variables and its controlled variables. Process decomposition is then realised after the screening of input and output variables. When the system decomposition is finished, the online subsystem modelling can be carried out by recursively block-wise renewing the samples. The proposed algorithm was applied in the Tennessee Eastman process and the validity was verified.
Modular Aero-Propulsion System Simulation
NASA Technical Reports Server (NTRS)
Parker, Khary I.; Guo, Ten-Huei
2006-01-01
The Modular Aero-Propulsion System Simulation (MAPSS) is a graphical simulation environment designed for the development of advanced control algorithms and rapid testing of these algorithms on a generic computational model of a turbofan engine and its control system. MAPSS is a nonlinear, non-real-time simulation comprising a Component Level Model (CLM) module and a Controller-and-Actuator Dynamics (CAD) module. The CLM module simulates the dynamics of engine components at a sampling rate of 2,500 Hz. The controller submodule of the CAD module simulates a digital controller, which has a typical update rate of 50 Hz. The sampling rate for the actuators in the CAD module is the same as that of the CLM. MAPSS provides a graphical user interface that affords easy access to engine-operation, engine-health, and control parameters; is used to enter such input model parameters as power lever angle (PLA), Mach number, and altitude; and can be used to change controller and engine parameters. Output variables are selectable by the user. Output data as well as any changes to constants and other parameters can be saved and reloaded into the GUI later.
A steering law for a roof-type configuration for a single-gimbal control moment gyro system
NASA Technical Reports Server (NTRS)
Yoshikawa, T.
1974-01-01
Single-Gimbal Control Moment Gyro (SGCMG) systems have been investigated for attitude control of the Large Space Telescope (LST) and the High Energy Astronomy Observatory (HEAO). However, various proposed steering laws for the SGCMG systems thus far have some defects because of singular states of the system. In this report, a steering law for a roof-type SGCMG system is proposed which is based on a new momentum distribution scheme that makes all the singular states unstable. This momentum distribution scheme is formulated by a treatment of the system as a sampled-data system. From analytical considerations, it is shown that this steering law gives control performance which is satisfactory for practical applications. Results of the preliminary computer simulation entirely support this premise.
Sampled-Data Consensus of Linear Multi-agent Systems With Packet Losses.
Zhang, Wenbing; Tang, Yang; Huang, Tingwen; Kurths, Jurgen
In this paper, the consensus problem is studied for a class of multi-agent systems with sampled data and packet losses, where random and deterministic packet losses are considered, respectively. For random packet losses, a Bernoulli-distributed white sequence is used to describe packet dropouts among agents in a stochastic way. For deterministic packet losses, a switched system with stable and unstable subsystems is employed to model packet dropouts in a deterministic way. The purpose of this paper is to derive consensus criteria, such that linear multi-agent systems with sampled-data and packet losses can reach consensus. By means of the Lyapunov function approach and the decomposition method, the design problem of a distributed controller is solved in terms of convex optimization. The interplay among the allowable bound of the sampling interval, the probability of random packet losses, and the rate of deterministic packet losses are explicitly derived to characterize consensus conditions. The obtained criteria are closely related to the maximum eigenvalue of the Laplacian matrix versus the second minimum eigenvalue of the Laplacian matrix, which reveals the intrinsic effect of communication topologies on consensus performance. Finally, simulations are given to show the effectiveness of the proposed results.In this paper, the consensus problem is studied for a class of multi-agent systems with sampled data and packet losses, where random and deterministic packet losses are considered, respectively. For random packet losses, a Bernoulli-distributed white sequence is used to describe packet dropouts among agents in a stochastic way. For deterministic packet losses, a switched system with stable and unstable subsystems is employed to model packet dropouts in a deterministic way. The purpose of this paper is to derive consensus criteria, such that linear multi-agent systems with sampled-data and packet losses can reach consensus. By means of the Lyapunov function approach and the decomposition method, the design problem of a distributed controller is solved in terms of convex optimization. The interplay among the allowable bound of the sampling interval, the probability of random packet losses, and the rate of deterministic packet losses are explicitly derived to characterize consensus conditions. The obtained criteria are closely related to the maximum eigenvalue of the Laplacian matrix versus the second minimum eigenvalue of the Laplacian matrix, which reveals the intrinsic effect of communication topologies on consensus performance. Finally, simulations are given to show the effectiveness of the proposed results.
Biomass Production System (BPS) plant growth unit.
Morrow, R C; Crabb, T M
2000-01-01
The Biomass Production System (BPS) was developed under the Small Business Innovative Research (SBIR) program to meet science, biotechnology and commercial plant growth needs in the Space Station era. The BPS is equivalent in size to a double middeck locker, but uses its own custom enclosure with a slide out structure to which internal components mount. The BPS contains four internal growth chambers, each with a growing volume of more than 4 liters. Each of the growth chambers has active nutrient delivery, and independent control of temperature, humidity, lighting, and CO2 set-points. Temperature control is achieved using a thermoelectric heat exchanger system. Humidity control is achieved using a heat exchanger with a porous interface which can both humidify and dehumidify. The control software utilizes fuzzy logic for nonlinear, coupled temperature and humidity control. The fluorescent lighting system can be dimmed to provide a range of light levels. CO2 levels are controlled by injecting pure CO2 to the system based on input from an infrared gas analyzer. The unit currently does not scrub CO2, but has been designed to accept scrubber cartridges. In addition to providing environmental control, a number of features are included to facilitate science. The BPS chambers are sealed to allow CO2 and water vapor exchange measurements. The plant chambers can be removed to allow manipulation or sampling of specimens, and each chamber has gas/fluid sample ports. A video camera is provided for each chamber, and frame-grabs and complete environmental data for all science and hardware system sensors are stored on an internal hard drive. Data files can also be transferred to 3.5-inch disks using the front panel disk drive.
Biomass Production System (BPS) Plant Growth Unit
NASA Astrophysics Data System (ADS)
Morrow, R. C.; Crabb, T. M.
The Biomass Production System (BPS) was developed under the Small Business Innovative Research (SBIR) program to meet science, biotechnology and commercial plant growth needs in the Space Station era. The BPS is equivalent in size to a double middeck locker, but uses it's own custom enclosure with a slide out structure to which internal components mount. The BPS contains four internal growth chambers, each with a growing volume of more than 4 liters. Each of the growth chambers has active nutrient delivery, and independent control of temperature, humidity, lighting, and CO2 set-points. Temperature control is achieved using a thermoelectric heat exchanger system. Humidity control is achieved using a heat exchanger with a porous interface which can both humidify and dehumidify. The control software utilizes fuzzy logic for nonlinear, coupled temperature and humidity control. The fluorescent lighting system can be dimmed to provide a range of light levels. CO2 levels are controlled by injecting pure CO2 to the system based on input from an infrared gas analyzer. The unit currently does not scrub CO2, but has been designed to accept scrubber cartridges. In addition to providing environmental control, a number of features are included to facilitate science. The BPS chambers are sealed to allow CO2 and water vapor exchange measurements. The plant chambers can be removed to allow manipulation or sampling of specimens, and each chamber has gas/fluid sample ports. A video camera is provided for each chamber, and frame-grabs and complete environmental data for all science and hardware system sensors are stored on an internal hard drive. Data files can also be transferred to 3.5-inch disks using the front panel disk drive
Pothineni, Sudhir Babu; Venugopalan, Nagarajan; Ogata, Craig M.; Hilgart, Mark C.; Stepanov, Sergey; Sanishvili, Ruslan; Becker, Michael; Winter, Graeme; Sauter, Nicholas K.; Smith, Janet L.; Fischetti, Robert F.
2014-01-01
The calculation of single- and multi-crystal data collection strategies and a data processing pipeline have been tightly integrated into the macromolecular crystallographic data acquisition and beamline control software JBluIce. Both tasks employ wrapper scripts around existing crystallographic software. JBluIce executes scripts through a distributed resource management system to make efficient use of all available computing resources through parallel processing. The JBluIce single-crystal data collection strategy feature uses a choice of strategy programs to help users rank sample crystals and collect data. The strategy results can be conveniently exported to a data collection run. The JBluIce multi-crystal strategy feature calculates a collection strategy to optimize coverage of reciprocal space in cases where incomplete data are available from previous samples. The JBluIce data processing runs simultaneously with data collection using a choice of data reduction wrappers for integration and scaling of newly collected data, with an option for merging with pre-existing data. Data are processed separately if collected from multiple sites on a crystal or from multiple crystals, then scaled and merged. Results from all strategy and processing calculations are displayed in relevant tabs of JBluIce. PMID:25484844
Pothineni, Sudhir Babu; Venugopalan, Nagarajan; Ogata, Craig M.; ...
2014-11-18
The calculation of single- and multi-crystal data collection strategies and a data processing pipeline have been tightly integrated into the macromolecular crystallographic data acquisition and beamline control software JBluIce. Both tasks employ wrapper scripts around existing crystallographic software. JBluIce executes scripts through a distributed resource management system to make efficient use of all available computing resources through parallel processing. The JBluIce single-crystal data collection strategy feature uses a choice of strategy programs to help users rank sample crystals and collect data. The strategy results can be conveniently exported to a data collection run. The JBluIce multi-crystal strategy feature calculates amore » collection strategy to optimize coverage of reciprocal space in cases where incomplete data are available from previous samples. The JBluIce data processing runs simultaneously with data collection using a choice of data reduction wrappers for integration and scaling of newly collected data, with an option for merging with pre-existing data. Data are processed separately if collected from multiple sites on a crystal or from multiple crystals, then scaled and merged. Results from all strategy and processing calculations are displayed in relevant tabs of JBluIce.« less
A Digital Control Algorithm for Magnetic Suspension Systems
NASA Technical Reports Server (NTRS)
Britton, Thomas C.
1996-01-01
An ongoing program exists to investigate and develop magnetic suspension technologies and modelling techniques at NASA Langley Research Center. Presently, there is a laboratory-scale large air-gap suspension system capable of five degree-of-freedom (DOF) control that is operational and a six DOF system that is under development. Those systems levitate a cylindrical element containing a permanent magnet core above a planar array of electromagnets, which are used for levitation and control purposes. In order to evaluate various control approaches with those systems, the Generic Real-Time State-Space Controller (GRTSSC) software package was developed. That control software package allows the user to implement multiple control methods and allows for varied input/output commands. The development of the control algorithm is presented. The desired functionality of the software is discussed, including the ability to inject noise on sensor inputs and/or actuator outputs. Various limitations, common issues, and trade-offs are discussed including data format precision; the drawbacks of using either Direct Memory Access (DMA), interrupts, or program control techniques for data acquisition; and platform dependent concerns related to the portability of the software, such as memory addressing formats. Efforts to minimize overall controller loop-rate and a comparison of achievable controller sample rates are discussed. The implementation of a modular code structure is presented. The format for the controller input data file and the noise information file is presented. Controller input vector information is available for post-processing by mathematical analysis software such as MATLAB1.
User`s manual for the CDC-1 digitizer controller
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ferron, J.R.
1994-09-01
A detailed description of how to use the CDC-1 digitizer controller is given. The CDC-1 is used with the CAMAC format digitizer models in the TRAQ series (manufactured by DSP Technology Inc.), the DAD-1 data acquisition daughter board, and the Intel i860-based SuperCard-2 (manufactured, by CSP Inc.) to form a high speed data acquisition and real time analysis system. Data can be transferred to the memory on the SuperCard-2 at a rate as high as 40 million 14-bit samples per second. Depending on the model of TRAQ digitizer in use, digitizing rates up to 3.33 MHz are supported (with eightmore » data channels), or, for instance, at a sample rate of 100 kHz, 384 data channels can be acquired.« less
NASA Technical Reports Server (NTRS)
Thau, F. E.; Montgomery, R. C.
1980-01-01
Techniques developed for the control of aircraft under changing operating conditions are used to develop a learning control system structure for a multi-configuration, flexible space vehicle. A configuration identification subsystem that is to be used with a learning algorithm and a memory and control process subsystem is developed. Adaptive gain adjustments can be achieved by this learning approach without prestoring of large blocks of parameter data and without dither signal inputs which will be suppressed during operations for which they are not compatible. The Space Shuttle Solar Electric Propulsion (SEP) experiment is used as a sample problem for the testing of adaptive/learning control system algorithms.
NASA Technical Reports Server (NTRS)
Sherman, W. L.
1975-01-01
The effects of steady wind, turbulence, data sample rate, and control-actuator natural frequency on the response of a possible automatic landing system were investigated in a nonstatistical study. The results indicate that the system, which interfaces with the microwave landing system, functions well in winds and turbulence as long as the guidance law contains proper compensation for wind. The system response was satisfactory down to five data samples per second, which makes the system compatible with the microwave landing system. No adverse effects were observed when actuator natural frequency was lowered. For limiting cases, those cases where the roll angle goes to zero just as the airplane touches down, the basic method for computing the turn-algorithm gains proved unsatisfactory and unacceptable landings resulted. Revised computation methods gave turn-algorithm gains that resulted in acceptable landings. The gains provided by the new method also improved the touchdown conditions for acceptable landings over those obtained when the gains were determined by the old method.
Integrated Seawater Sampler and Data Acquisition System Prototype
1993-04-01
circuitry. The sample bags, initially sealed in a chemically clean environment , were opened by pumping the water out of the tray, thus forcing water into...the water sample containers from materials having low metals content. Control of the sample transfer environment through the use of Class-100 clean... environment . CTD designers also intended their underwater units and deck terminals to be connected directly together providing AC grounding appropriate to
Identification and Quantitative Measurements of Chemical Species by Mass Spectrometry
NASA Technical Reports Server (NTRS)
Zondlo, Mark A.; Bomse, David S.
2005-01-01
The development of a miniature gas chromatograph/mass spectrometer system for the measurement of chemical species of interest to combustion is described. The completed system is a fully-contained, automated instrument consisting of a sampling inlet, a small-scale gas chromatograph, a miniature, quadrupole mass spectrometer, vacuum pumps, and software. A pair of computer-driven valves controls the gas sampling and introduction to the chromatographic column. The column has a stainless steel exterior and a silica interior, and contains an adsorbent of that is used to separate organic species. The detection system is based on a quadrupole mass spectrometer consisting of a micropole array, electrometer, and a computer interface. The vacuum system has two miniature pumps to maintain the low pressure needed for the mass spectrometer. A laptop computer uses custom software to control the entire system and collect the data. In a laboratory demonstration, the system separated calibration mixtures containing 1000 ppm of alkanes and alkenes.
The EnzymeTracker: an open-source laboratory information management system for sample tracking.
Triplet, Thomas; Butler, Gregory
2012-01-26
In many laboratories, researchers store experimental data on their own workstation using spreadsheets. However, this approach poses a number of problems, ranging from sharing issues to inefficient data-mining. Standard spreadsheets are also error-prone, as data do not undergo any validation process. To overcome spreadsheets inherent limitations, a number of proprietary systems have been developed, which laboratories need to pay expensive license fees for. Those costs are usually prohibitive for most laboratories and prevent scientists from benefiting from more sophisticated data management systems. In this paper, we propose the EnzymeTracker, a web-based laboratory information management system for sample tracking, as an open-source and flexible alternative that aims at facilitating entry, mining and sharing of experimental biological data. The EnzymeTracker features online spreadsheets and tools for monitoring numerous experiments conducted by several collaborators to identify and characterize samples. It also provides libraries of shared data such as protocols, and administration tools for data access control using OpenID and user/team management. Our system relies on a database management system for efficient data indexing and management and a user-friendly AJAX interface that can be accessed over the Internet. The EnzymeTracker facilitates data entry by dynamically suggesting entries and providing smart data-mining tools to effectively retrieve data. Our system features a number of tools to visualize and annotate experimental data, and export highly customizable reports. It also supports QR matrix barcoding to facilitate sample tracking. The EnzymeTracker was designed to be easy to use and offers many benefits over spreadsheets, thus presenting the characteristics required to facilitate acceptance by the scientific community. It has been successfully used for 20 months on a daily basis by over 50 scientists. The EnzymeTracker is freely available online at http://cubique.fungalgenomics.ca/enzymedb/index.html under the GNU GPLv3 license.
The EnzymeTracker: an open-source laboratory information management system for sample tracking
2012-01-01
Background In many laboratories, researchers store experimental data on their own workstation using spreadsheets. However, this approach poses a number of problems, ranging from sharing issues to inefficient data-mining. Standard spreadsheets are also error-prone, as data do not undergo any validation process. To overcome spreadsheets inherent limitations, a number of proprietary systems have been developed, which laboratories need to pay expensive license fees for. Those costs are usually prohibitive for most laboratories and prevent scientists from benefiting from more sophisticated data management systems. Results In this paper, we propose the EnzymeTracker, a web-based laboratory information management system for sample tracking, as an open-source and flexible alternative that aims at facilitating entry, mining and sharing of experimental biological data. The EnzymeTracker features online spreadsheets and tools for monitoring numerous experiments conducted by several collaborators to identify and characterize samples. It also provides libraries of shared data such as protocols, and administration tools for data access control using OpenID and user/team management. Our system relies on a database management system for efficient data indexing and management and a user-friendly AJAX interface that can be accessed over the Internet. The EnzymeTracker facilitates data entry by dynamically suggesting entries and providing smart data-mining tools to effectively retrieve data. Our system features a number of tools to visualize and annotate experimental data, and export highly customizable reports. It also supports QR matrix barcoding to facilitate sample tracking. Conclusions The EnzymeTracker was designed to be easy to use and offers many benefits over spreadsheets, thus presenting the characteristics required to facilitate acceptance by the scientific community. It has been successfully used for 20 months on a daily basis by over 50 scientists. The EnzymeTracker is freely available online at http://cubique.fungalgenomics.ca/enzymedb/index.html under the GNU GPLv3 license. PMID:22280360
Hydrologic data for urban stormwater studies in the Dallas-Fort Worth area, Texas, 1992-94
Baldys, Stanley; Raines, T.H.; Mansfield, B.L.; Sandlin, J.T.
1997-01-01
This report presents precipitation and waterquality data from analyses of 210 samples collected at 30 storm-sewer outfall stations in the Dallas-Fort Worth area, Texas, during February 1992-November 1994. The data were collected to fulfill requirements mandated by the U.S. Environmental Protection Agency to the cities of Arlington, Dallas, Fort Worth, Garland, Irving, Mesquite, and Piano and to the Dallas and Fort Worth Districts of the Texas Department of Transportation to obtain a National Pollution Discharge Elimination System permit. Data were collected at storm-sewer outfall stations in drainage basins classified as singular land use, either residential, commercial, industrial, or highway. Also included are qualityassurance/quality-control data for samples collected in conjunction with the stormwater samples.
1984-02-01
measurable impact if changed. The following items were included in the sample: * Mark Zero Items -Low demand insurance items which represent about three...R&D efforts reviewed. The resulting assessment highlighted the generic enabling technologies and cross- cutting R&D projects required to focus current...supplied by spot buys, and which may generate Navy Inventory Control Numbers (NICN). Random samples of data were extracted from the Master Data File ( MDF
Next generation control system for reflexive aerostructures
NASA Astrophysics Data System (ADS)
Maddux, Michael R.; Meents, Elizabeth P.; Barnell, Thomas J.; Cable, Kristin M.; Hemmelgarn, Christopher; Margraf, Thomas W.; Havens, Ernie
2010-04-01
Cornerstone Research Group Inc. (CRG) has developed and demonstrated a composite structural solution called reflexive composites for aerospace applications featuring CRG's healable shape memory polymer (SMP) matrix. In reflexive composites, an integrated structural health monitoring (SHM) system autonomously monitors the structural health of composite aerospace structures, while integrated intelligent controls monitor data from the SHM system to characterize damage and initiate healing when damage is detected. Development of next generation intelligent controls for reflexive composites were initiated for the purpose of integrating prognostic health monitoring capabilities into the reflexive composite structural solution. Initial efforts involved data generation through physical inspections and mechanical testing. Compression after impact (CAI) testing was conducted on composite-reinforced shape memory polymer samples to induce damage and investigate the effectiveness of matrix healing on mechanical performance. Non-destructive evaluation (NDE) techniques were employed to observe and characterize material damage. Restoration of mechanical performance was demonstrated through healing, while NDE data showed location and size of damage and verified mitigation of damage post-healing. Data generated was used in the development of next generation reflexive controls software. Data output from the intelligent controls could serve as input to Integrated Vehicle Health Management (IVHM) systems and Integrated Resilient Aircraft Controls (IRAC). Reflexive composite technology has the ability to reduce maintenance required on composite structures through healing, offering potential to significantly extend service life of aerospace vehicles and reduce operating and lifecycle costs.
Garrison, R A; Robertson, L D; Koehn, R D; Wynn, S R
1993-12-01
Commercial air duct sanitation services are advertised to the public as being effective in reducing indoor aeroallergen levels despite the absence of published supporting data. Eight residential heat-ventilation-air conditioning (HVAC) systems in six homes and seven HVAC systems in five homes in winter and summer, respectively, were sampled to determine fungal colony forming units (CFUs) prior to and after an HVAC sanitation procedure was performed by a local company. Two houses in which no sanitation procedure was performed served as controls in each study phase. Two sample sets were obtained at each HVAC system prior to cleaning in order to determine baseline CFU levels. The test HVAC systems were then cleaned, and the HVAC systems allowed to operate as desired by the residents. Posttreatment sampling was performed 48 hours and then weekly after cleaning for 8 weeks. The HVAC systems were analyzed by exposing sterile 2% malt extract media plates at a 90-degree angle to the air flow at the air supply and air return vents. The baseline CFUs were similar in the control and study houses. Eight weeks after sanitation, the study houses demonstrated an overall CFU reduction of 92% during winter and 84% during summer. No reduction in CFU values was observed over the 8-week study period for the houses selected as controls. Further, HVAC sanitation appeared to reduce the number of fungal colonies entering and leaving the HVAC system, suggesting that the HVAC contained a significant percentage of the total fungal load in these homes. These data suggest that HVAC sanitation may be an effective tool in reducing airborne fungal populations in residential environments.
Ueki, Shigeharu; Kayaba, Hiroyuki; Tomita, Noriko; Kobayashi, Noriko; Takahashi, Tomoe; Obara, Toshikage; Takeda, Masahide; Moritoki, Yuki; Itoga, Masamichi; Ito, Wataru; Ohsaga, Atsushi; Kondoh, Katsuyuki; Chihara, Junichi
2011-04-01
The active involvement of hospital laboratory in surveillance is crucial to the success of nosocomial infection control. The recent dramatic increase of antimicrobial-resistant organisms and their spread into the community suggest that the infection control strategy of independent medical institutions is insufficient. To share the clinical data and surveillance in our local medical region, we developed a microbiology data warehouse for networking hospital laboratories in Akita prefecture. This system, named Akita-ReNICS, is an easy-to-use information management system designed to compare, track, and report the occurrence of antimicrobial-resistant organisms. Participating laboratories routinely transfer their coded and formatted microbiology data to ReNICS server located at Akita University Hospital from their health care system's clinical computer applications over the internet. We established the system to automate the statistical processes, so that the participants can access the server to monitor graphical data in the manner they prefer, using their own computer's browser. Furthermore, our system also provides the documents server, microbiology and antimicrobiotic database, and space for long-term storage of microbiological samples. Akita-ReNICS could be a next generation network for quality improvement of infection control.
Asynchronous sampled-data approach for event-triggered systems
NASA Astrophysics Data System (ADS)
Mahmoud, Magdi S.; Memon, Azhar M.
2017-11-01
While aperiodically triggered network control systems save a considerable amount of communication bandwidth, they also pose challenges such as coupling between control and event-condition design, optimisation of the available resources such as control, communication and computation power, and time-delays due to computation and communication network. With this motivation, the paper presents separate designs of control and event-triggering mechanism, thus simplifying the overall analysis, asynchronous linear quadratic Gaussian controller which tackles delays and aperiodic nature of transmissions, and a novel event mechanism which compares the cost of the aperiodic system against a reference periodic implementation. The proposed scheme is simulated on a linearised wind turbine model for pitch angle control and the results show significant improvement against the periodic counterpart.
Space Station CMIF extended duration metabolic control test
NASA Technical Reports Server (NTRS)
Schunk, Richard G.; Bagdigian, Robert M.; Carrasquillo, Robyn L.; Ogle, Kathryn Y.; Wieland, Paul O.
1989-01-01
The Space Station Extended Duration Metabolic Control Test (EMCT) was conducted at the MSFC Core Module Integration Facility. The primary objective of the EMCT was to gather performance data from a partially-closed regenerative Environmental Control and Life Support (ECLS) system functioning under steady-state conditions. Included is a description of the EMCT configuration, a summary of events, a discussion of anomalies that occurred during the test, and detailed results and analysis from individual measurements of water and gas samples taken during the test. A comparison of the physical, chemical, and microbiological methods used in the post test laboratory analyses of the water samples is included. The preprototype ECLS hardware used in the test, providing an overall process description and theory of operation for each hardware item. Analytical results pertaining to a system level mass balance and selected system power estimates are also included.
Fuzzy Adaptive Control Design and Discretization for a Class of Nonlinear Uncertain Systems.
Zhao, Xudong; Shi, Peng; Zheng, Xiaolong
2016-06-01
In this paper, tracking control problems are investigated for a class of uncertain nonlinear systems in lower triangular form. First, a state-feedback controller is designed by using adaptive backstepping technique and the universal approximation ability of fuzzy logic systems. During the design procedure, a developed method with less computation is proposed by constructing one maximum adaptive parameter. Furthermore, adaptive controllers with nonsymmetric dead-zone are also designed for the systems. Then, a sampled-data control scheme is presented to discretize the obtained continuous-time controller by using the forward Euler method. It is shown that both proposed continuous and discrete controllers can ensure that the system output tracks the target signal with a small bounded error and the other closed-loop signals remain bounded. Two simulation examples are presented to verify the effectiveness and applicability of the proposed new design techniques.
Thermostatic system of sensor in NIR spectrometer based on PID control
NASA Astrophysics Data System (ADS)
Wang, Zhihong; Qiao, Liwei; Ji, Xufei
2016-11-01
Aiming at the shortcomings of the primary sensor thermostatic control system in the near infrared (NIR) spectrometer, a novel thermostatic control system based on proportional-integral-derivative (PID) control technology was developed to improve the detection precision of the NIR spectrometer. There were five parts including bridge amplifier circuit, analog-digital conversion (ADC) circuit, microcontroller, digital-analog conversion (DAC) circuit and drive circuit in the system. The five parts formed a closed-loop control system based on PID algorithm that was used to control the error between the temperature calculated by the sampling data of ADC and the designed temperature to ensure the stability of the spectrometer's sensor. The experimental results show that, when the operating temperature of sensor is -11°, compared with the original system, the temperature control precision of the new control system is improved from ±0.64° to ±0.04° and the spectrum signal to noise ratio (SNR) is improved from 4891 to 5967.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Estill, C.F.; Kovein, R.J.; Jones, J.H.
1999-03-26
The National Institute for Occupational Safety and Health (NIOSH) is currently conducting research on ventilation controls to reduce furniture stripping exposures to methylene chloride to the OSHA PEL of 25 ppm. Low cost ventilation systems were designed by NIOSH researchers along with Benny Bixenman of Benco Sales, Inc. (Forney, TX). The controls were constructed and installed by Benco Sales. This report compares the methylene chloride levels of one worker stripping furniture using the recently installed ventilation controls and using the existing controls. During the survey, two different chemical stripping solutions (a standard formulation and a low methylene chloride content formulation)more » were used and compared. This survey tested three control combinations: (1) new ventilation, low methylene chloride stripper, (2) new ventilation, standard stripping solution, and (3) old ventilation, standard stripping solution. During each test, sorbent tube sampling and real-time sampling were employed. Sorbent tube, data collected in the worker's breathing zone, ranged from 300 to 387 ppm. Real-time data showed breathing zone exposures to range from 211 to 383 ppm while stripping and 164 to 230 ppm while rinsing. Data were inconclusive to determine which ventilation system or stripping solution produced the lowest exposures. Recommendations are made in the report to improve the newly installed ventilation controls.« less
Luján, J L; Crago, P E
2004-11-01
Neuroprosthestic systems can be used to restore hand grasp and wrist control in individuals with C5/C6 spinal cord injury. A computer-based system was developed for the implementation, tuning and clinical assessment of neuroprosthetic controllers, using off-the-shelf hardware and software. The computer system turned a Pentium III PC running Windows NT into a non-dedicated, real-time system for the control of neuroprostheses. Software execution (written using the high-level programming languages LabVIEW and MATLAB) was divided into two phases: training and real-time control. During the training phase, the computer system collected input/output data by stimulating the muscles and measuring the muscle outputs in real-time, analysed the recorded data, generated a set of training data and trained an artificial neural network (ANN)-based controller. During real-time control, the computer system stimulated the muscles using stimulus pulsewidths predicted by the ANN controller in response to a sampled input from an external command source, to provide independent control of hand grasp and wrist posture. System timing was stable, reliable and capable of providing muscle stimulation at frequencies up to 24Hz. To demonstrate the application of the test-bed, an ANN-based controller was implemented with three inputs and two independent channels of stimulation. The ANN controller's ability to control hand grasp and wrist angle independently was assessed by quantitative comparison of the outputs of the stimulated muscles with a set of desired grasp or wrist postures determined by the command signal. Controller performance results were mixed, but the platform provided the tools to implement and assess future controller designs.
Development of a sensitivity analysis technique for multiloop flight control systems
NASA Technical Reports Server (NTRS)
Vaillard, A. H.; Paduano, J.; Downing, D. R.
1985-01-01
This report presents the development and application of a sensitivity analysis technique for multiloop flight control systems. This analysis yields very useful information on the sensitivity of the relative-stability criteria of the control system, with variations or uncertainties in the system and controller elements. The sensitivity analysis technique developed is based on the computation of the singular values and singular-value gradients of a feedback-control system. The method is applicable to single-input/single-output as well as multiloop continuous-control systems. Application to sampled-data systems is also explored. The sensitivity analysis technique was applied to a continuous yaw/roll damper stability augmentation system of a typical business jet, and the results show that the analysis is very useful in determining the system elements which have the largest effect on the relative stability of the closed-loop system. As a secondary product of the research reported here, the relative stability criteria based on the concept of singular values were explored.
Design of a portable electronic nose for real-fake detection of liquors
NASA Astrophysics Data System (ADS)
Qi, Pei-Feng; Zeng, Ming; Li, Zhi-Hua; Sun, Biao; Meng, Qing-Hao
2017-09-01
Portability is a major issue that influences the practical application of electronic noses (e-noses). For liquors detection, an e-nose must preprocess the liquid samples (e.g., using evaporation and thermal desorption), which makes the portable design even more difficult. To realize convenient and rapid detection of liquors, we designed a portable e-nose platform that consists of hardware and software systems. The hardware system contains an evaporation/sampling module, a reaction module, a control/data acquisition and analysis module, and a power module. The software system provides a user-friendly interface and can achieve automatic sampling and data processing. This e-nose platform has been applied to the real-fake recognition of Chinese liquors. Through parameter optimization of a one-class support vector machine classifier, the error rate of the negative samples is greatly reduced, and the overall recognition accuracy is improved. The results validated the feasibility of the designed portable e-nose platform.
CFO compensation method using optical feedback path for coherent optical OFDM system
NASA Astrophysics Data System (ADS)
Moon, Sang-Rok; Hwang, In-Ki; Kang, Hun-Sik; Chang, Sun Hyok; Lee, Seung-Woo; Lee, Joon Ki
2017-07-01
We investigate feasibility of carrier frequency offset (CFO) compensation method using optical feedback path for coherent optical orthogonal frequency division multiplexing (CO-OFDM) system. Recently proposed CFO compensation algorithms provide wide CFO estimation range in electrical domain. However, their practical compensation range is limited by sampling rate of an analog-to-digital converter (ADC). This limitation has not drawn attention, since the ADC sampling rate was high enough comparing to the data bandwidth and CFO in the wireless OFDM system. For CO-OFDM, the limitation is becoming visible because of increased data bandwidth, laser instability (i.e. large CFO) and insufficient ADC sampling rate owing to high cost. To solve the problem and extend practical CFO compensation range, we propose a CFO compensation method having optical feedback path. By adding simple wavelength control for local oscillator, the practical CFO compensation range can be extended to the sampling frequency range. The feasibility of the proposed method is experimentally investigated.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oden, L.L.; O`Conner, W.K.; Turner, P.C.
1993-11-19
This report presents field results and raw data from the Buried Waste Integrated Demonstration (BWID) Arc Melter Vitrification Project Phase 1 baseline test series conducted by the Idaho National Engineering Laboratory (INEL) in cooperation with the U.S. Bureau of Mines (USBM). The baseline test series was conducted using the electric arc melter facility at the USBM Albany Research Center in Albany, Oregon. Five different surrogate waste feed mixtures were tested that simulated thermally-oxidized, buried, TRU-contaminated, mixed wastes and soils present at the INEL. The USBM Arc Furnace Integrated Waste Processing Test Facility includes a continuous feed system, the arc meltingmore » furnace, an offgas control system, and utilities. The melter is a sealed, 3-phase alternating current (ac) furnace approximately 2 m high and 1.3 m wide. The furnace has a capacity of 1 metric ton of steel and can process as much as 1,500 lb/h of soil-type waste materials. The surrogate feed materials included five mixtures designed to simulate incinerated TRU-contaminated buried waste materials mixed with INEL soil. Process samples, melter system operations data and offgas composition data were obtained during the baseline tests to evaluate the melter performance and meet test objectives. Samples and data gathered during this program included (a) automatically and manually logged melter systems operations data, (b) process samples of slag, metal and fume solids, and (c) offgas composition, temperature, velocity, flowrate, moisture content, particulate loading and metals content. This report consists of 2 volumes: Volume I summarizes the baseline test operations. It includes an executive summary, system and facility description, review of the surrogate waste mixtures, and a description of the baseline test activities, measurements, and sample collection. Volume II contains the raw test data and sample analyses from samples collected during the baseline tests.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oden, L.L.; O`Connor, W.K.; Turner, P.C.
1993-11-19
This report presents field results and raw data from the Buried Waste Integrated Demonstration (BWID) Arc Melter Vitrification Project Phase 1 baseline test series conducted by the Idaho National Engineering Laboratory (INEL) in cooperation with the U.S. Bureau of Mines (USBM). The baseline test series was conducted using the electric arc melter facility at the USBM Albany Research Center in Albany, Oregon. Five different surrogate waste feed mixtures were tested that simulated thermally-oxidized, buried, TRU-contaminated, mixed wastes and soils present at the INEL. The USBM Arc Furnace Integrated Waste Processing Test Facility includes a continuous feed system, the arc meltingmore » furnace, an offgas control system, and utilities. The melter is a sealed, 3-phase alternating current (ac) furnace approximately 2 m high and 1.3 m wide. The furnace has a capacity of 1 metric ton of steel and can process as much as 1,500 lb/h of soil-type waste materials. The surrogate feed materials included five mixtures designed to simulate incinerated TRU-contaminated buried waste materials mixed with INEL soil. Process samples, melter system operations data and offgas composition data were obtained during the baseline tests to evaluate the melter performance and meet test objectives. Samples and data gathered during this program included (a) automatically and manually logged melter systems operations data, (b) process samples of slag, metal and fume solids, and (c) offgas composition, temperature, velocity, flowrate, moisture content, particulate loading and metals content. This report consists of 2 volumes: Volume I summarizes the baseline test operations. It includes an executive summary, system and facility description, review of the surrogate waste mixtures, and a description of the baseline test activities, measurements, and sample collection. Volume II contains the raw test data and sample analyses from samples collected during the baseline tests.« less
Hazard Control Extensions in a COTS Based Data Handling System
NASA Astrophysics Data System (ADS)
Vogel, Torsten; Rakers, Sven; Gronowski, Matthias; Schneegans, Joachim
2011-08-01
EML is an electromagnetic levitator for containerless processing of conductive samples on the International Space Station. This material sciences experiment is running in the European Drawer Rack (EDR) facility. The objective of this experiment is to gain insight into the parameters of liquid metal samples and their crystallisation processes without the influence of container walls. To this end the samples are electromagnetically positioned in a coil system and then heated up beyond their melting point in an ultraclean environment.The EML programme is currently under development by Astrium Space Transportation in Friedrichshafen and Bremen; jointly funded by ESA and DLR (on behalf of BMWi, contract 50WP0808). EML consists of four main modules listed in Table 1. The paper focuses mainly on the architecture and design of the ECM module and its contribution to a safe operation of the experiment. The ECM is a computer system that integrates the power supply to the EML experiment, control functions and video handling and compression features. Experiment control is performed by either telecommand or the execution of predefined experiment scripts.
NASA Technical Reports Server (NTRS)
Mitchell, J. R.
1972-01-01
The frequency response method of analyzing control system performance is discussed, and the difficulty of obtaining the sampled frequency response of the continuous system is considered. An upper bound magnitude error equation is obtained which yields reasonable estimates of the actual error. Finalization of the compensator improvement program is also reported, and the program was used to design compensators for Saturn 5/S1-C dry workshop and Saturn 5/S1-C Skylab.
A dedicated database system for handling multi-level data in systems biology.
Pornputtapong, Natapol; Wanichthanarak, Kwanjeera; Nilsson, Avlant; Nookaew, Intawat; Nielsen, Jens
2014-01-01
Advances in high-throughput technologies have enabled extensive generation of multi-level omics data. These data are crucial for systems biology research, though they are complex, heterogeneous, highly dynamic, incomplete and distributed among public databases. This leads to difficulties in data accessibility and often results in errors when data are merged and integrated from varied resources. Therefore, integration and management of systems biological data remain very challenging. To overcome this, we designed and developed a dedicated database system that can serve and solve the vital issues in data management and hereby facilitate data integration, modeling and analysis in systems biology within a sole database. In addition, a yeast data repository was implemented as an integrated database environment which is operated by the database system. Two applications were implemented to demonstrate extensibility and utilization of the system. Both illustrate how the user can access the database via the web query function and implemented scripts. These scripts are specific for two sample cases: 1) Detecting the pheromone pathway in protein interaction networks; and 2) Finding metabolic reactions regulated by Snf1 kinase. In this study we present the design of database system which offers an extensible environment to efficiently capture the majority of biological entities and relations encountered in systems biology. Critical functions and control processes were designed and implemented to ensure consistent, efficient, secure and reliable transactions. The two sample cases on the yeast integrated data clearly demonstrate the value of a sole database environment for systems biology research.
Isolated thermocouple amplifier system for stirred fixed-bed gasifier
Fasching, George E.
1992-01-01
A sensing system is provided for determining the bed temperature profile of the bed of a stirred, fixed-bed gasifier including a plurality of temperature sensors for sensing the bed temperature at different levels, a transmitter for transmitting data based on the outputs of the sensors to a remote operator's station, and a battery-based power supply. The system includes an isolation amplifier system comprising a plurality of isolation amplifier circuits for amplifying the outputs of the individual sensors. The isolation amplifier circuits each comprise an isolation operational amplifier connected to a sensor; a first "flying capacitor" circuit for, in operation, controlling the application of power from the power supply to the isolation amplifier; an output sample and hold circuit connected to the transmitter; a second "flying capacitor" circuit for, in operation, controlling the transfer of the output of the isolation amplifier to the sample and hold circuit; and a timing and control circuit for activating the first and second capacitor circuits in a predetermined timed sequence.
NASA Astrophysics Data System (ADS)
Robson, E. I.; Stevens, J. A.; Jenness, T.
2001-11-01
Calibrated data for 65 flat-spectrum extragalactic radio sources are presented at a wavelength of 850μm, covering a three-year period from 1997 April. The data, obtained from the James Clerk Maxwell Telescope using the SCUBA camera in pointing mode, were analysed using an automated pipeline process based on the Observatory Reduction and Acquisition Control-Data Reduction (orac-dr) system. This paper describes the techniques used to analyse and calibrate the data, and presents the data base of results along with a representative sample of the better-sampled light curves.
Mars oxygen production system design
NASA Technical Reports Server (NTRS)
Cotton, Charles E.; Pillow, Linda K.; Perkinson, Robert C.; Brownlie, R. P.; Chwalowski, P.; Carmona, M. F.; Coopersmith, J. P.; Goff, J. C.; Harvey, L. L.; Kovacs, L. A.
1989-01-01
The design and construction phase is summarized of the Mars oxygen demonstration project. The basic hardware required to produce oxygen from simulated Mars atmosphere was assembled and tested. Some design problems still remain with the sample collection and storage system. In addition, design and development of computer compatible data acquisition and control instrumentation is ongoing.
Mars oxygen production system design
NASA Technical Reports Server (NTRS)
1988-01-01
This report summarizes the design and construction of the Mars oxygen demonstration project. The basic hardware required to produce oxygen from simulated Mars atmosphere has been assembled and tested. Some design problems still remain with the sample collection and storage system. In addition, design and development of computer data acquisition and control instrumentation is continuing.
Wildt, Signe; Krag, Aleksander; Gluud, Liselotte
2011-01-01
Objectives To evaluate the adequacy of reporting of protocols for randomised trials on diseases of the digestive system registered in http://ClinicalTrials.gov and the consistency between primary outcomes, secondary outcomes and sample size specified in http://ClinicalTrials.gov and published trials. Methods Randomised phase III trials on adult patients with gastrointestinal diseases registered before January 2009 in http://ClinicalTrials.gov were eligible for inclusion. From http://ClinicalTrials.gov all data elements in the database required by the International Committee of Medical Journal Editors (ICMJE) member journals were extracted. The subsequent publications for registered trials were identified. For published trials, data concerning publication date, primary and secondary endpoint, sample size, and whether the journal adhered to ICMJE principles were extracted. Differences between primary and secondary outcomes, sample size and sample size calculations data in http://ClinicalTrials.gov and in the published paper were registered. Results 105 trials were evaluated. 66 trials (63%) were published. 30% of trials were registered incorrectly after their completion date. Several data elements of the required ICMJE data list were not filled in, with missing data in 22% and 11%, respectively, of cases concerning the primary outcome measure and sample size. In 26% of the published papers, data on sample size calculations were missing and discrepancies between sample size reporting in http://ClinicalTrials.gov and published trials existed. Conclusion The quality of registration of randomised controlled trials still needs improvement.
NASA Astrophysics Data System (ADS)
Zhao, Liyun; Zhou, Jin; Wu, Quanjun
2016-01-01
This paper considers the sampled-data synchronisation problems of coupled harmonic oscillators with communication and input delays subject to controller failure. A synchronisation protocol is proposed for such oscillator systems over directed network topology, and then some general algebraic criteria on exponential convergence for the proposed protocol are established. The main features of the present investigation include: (1) both the communication and input delays are simultaneously addressed, and the directed network topology is firstly considered and (2) the effects of time delays on synchronisation performance are theoretically and numerically investigated. It is shown that in the absence of communication delays, coupled harmonic oscillators can achieve synchronisation oscillatory motion. Whereas if communication delays are nonzero at infinite multiple sampled-data instants, its synchronisation (or consensus) state is zero. This conclusion can be used as an effective control strategy to stabilise coupled harmonic oscillators in practical applications. Furthermore, it is interesting to find that increasing either communication or input delays will enhance the synchronisation performance of coupled harmonic oscillators. Subsequently, numerical examples illustrate and visualise theoretical results.
Holmes, Thomas D; Guilmette, Raymond A; Cheng, Yung Sung; Parkhurst, Mary Ann; Hoover, Mark D
2009-03-01
The Capstone Depleted Uranium (DU) Aerosol Study was undertaken to obtain aerosol samples resulting from a large-caliber DU penetrator striking an Abrams or Bradley test vehicle. The sampling strategy was designed to (1) optimize the performance of the samplers and maintain their integrity in the extreme environment created during perforation of an armored vehicle by a DU penetrator, (2) collect aerosols as a function of time post perforation, and (3) obtain size-classified samples for analysis of chemical composition, particle morphology, and solubility in lung fluid. This paper describes the experimental setup and sampling methodologies used to achieve these objectives. Custom-designed arrays of sampling heads were secured to the inside of the target in locations approximating the breathing zones of the crew locations in the test vehicles. Each array was designed to support nine filter cassettes and nine cascade impactors mounted with quick-disconnect fittings. Shielding and sampler placement strategies were used to minimize sampler loss caused by the penetrator impact and the resulting fragments of eroded penetrator and perforated armor. A cyclone train was used to collect larger quantities of DU aerosol for measurement of chemical composition and solubility. A moving filter sample was used to obtain semicontinuous samples for DU concentration determination. Control for the air samplers was provided by five remotely located valve control and pressure monitoring units located inside and around the test vehicle. These units were connected to a computer interface chassis and controlled using a customized LabVIEW engineering computer control program. The aerosol sampling arrays and control systems for the Capstone study provided the needed aerosol samples for physicochemical analysis, and the resultant data were used for risk assessment of exposure to DU aerosol.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Holmes, Thomas D.; Guilmette, Raymond A.; Cheng, Yung-Sung
2009-03-01
The Capstone Depleted Uranium Aerosol Study was undertaken to obtain aerosol samples resulting from a kinetic-energy cartridge with a large-caliber depleted uranium (DU) penetrator striking an Abrams or Bradley test vehicle. The sampling strategy was designed to (1) optimize the performance of the samplers and maintain their integrity in the extreme environment created during perforation of an armored vehicle by a DU penetrator, (2) collect aerosols as a function of time post-impact, and (3) obtain size-classified samples for analysis of chemical composition, particle morphology, and solubility in lung fluid. This paper describes the experimental setup and sampling methodologies used tomore » achieve these objectives. Custom-designed arrays of sampling heads were secured to the inside of the target in locations approximating the breathing zones of the vehicle commander, loader, gunner, and driver. Each array was designed to support nine filter cassettes and nine cascade impactors mounted with quick-disconnect fittings. Shielding and sampler placement strategies were used to minimize sampler loss caused by the penetrator impact and the resulting fragments of eroded penetrator and perforated armor. A cyclone train was used to collect larger quantities of DU aerosol for chemical composition and solubility. A moving filter sample was used to obtain semicontinuous samples for depleted uranium concentration determination. Control for the air samplers was provided by five remotely located valve control and pressure monitoring units located inside and around the test vehicle. These units were connected to a computer interface chassis and controlled using a customized LabVIEW engineering computer control program. The aerosol sampling arrays and control systems for the Capstone study provided the needed aerosol samples for physicochemical analysis, and the resultant data were used for risk assessment of exposure to DU aerosol.« less
Vierling, L.A.; Fersdahl, M.; Chen, X.; Li, Z.; Zimmerman, P.
2006-01-01
We describe a new remote sensing system called the Short Wave Aerostat-Mounted Imager (SWAMI). The SWAMI is designed to acquire co-located video imagery and hyperspectral data to study basic remote sensing questions and to link landscape level trace gas fluxes with spatially and temporally appropriate spectral observations. The SWAMI can fly at altitudes up to 2 km above ground level to bridge the spatial gap between radiometric measurements collected near the surface and those acquired by other aircraft or satellites. The SWAMI platform consists of a dual channel hyperspectral spectroradiometer, video camera, GPS, thermal infrared sensor, and several meteorological and control sensors. All SWAMI functions (e.g. data acquisition and sensor pointing) can be controlled from the ground via wireless transmission. Sample data from the sampling platform are presented, along with several potential scientific applications of SWAMI data.
Health Services Utilization among Children with and without Autism Spectrum Disorders
ERIC Educational Resources Information Center
Cummings, Janet R.; Lynch, Frances L.; Rust, Kristal C.; Coleman, Karen J.; Madden, Jeanne M.; Owen-Smith, Ashli A.; Yau, Vincent M.; Qian, Yinge; Pearson, Kathryn A.; Crawford, Phillip M.; Massolo, Maria L.; Quinn, Virginia P.; Croen, Lisa A.
2016-01-01
Using data from multiple health systems (2009-2010) and the largest sample to date, this study compares health services use among youth with and without an autism spectrum disorder (ASD)--including preventive services not previously studied. To examine these differences, we estimated logistic and count data models, controlling for demographic…
From field to cloud: a collaborative software tool to manage hydrological observatories
NASA Astrophysics Data System (ADS)
Kraft, Philipp; Weber, Chris P.; Windhorst, David; Breuer, Lutz
2017-04-01
Managing data collection, assessment, storage, and analysis in hydrological observatories is challenging: Many processes can only be detected when long-term time series are being analysed, but temporary staff like postgraduates perform the measurements. Naturally the students focus on the data needed for their project and do not particularly care about the long-term availability of the data. Data providing new process insights gets often lost in unmaintainable spreadsheets with no clear distinction between raw, error controlled and derived data. Data warehouse systems, like the one developed by the Consortium of Universities for the Advancement of Hydrologic Science (CUAHSI) and strict data management guide lines by funding institutions, intend to mediate this effect. However, data warehouse structures are optimized for write once / read often use and require rigorous quality control and metadata description prior to the upload. Our experience shows a risk for data loss at this stage: Data collected at the end of a project is not reviewed and never enters the database and gets lost with the expiring position. As a solution to this kind of problems, we suggest to enter observation early, if possible online, and perform the review process in the system. We are presenting a new collaborative tool for managing hydrological observatories in a standardized and well documented manner directly from the point of data production, the field. Beside the observation data the system stores the observatory management tasks to ensure regular sampling or sensor maintenance. A second benefit of logging management actions together with observations, is the possibility to interpret side effects of sampling or maintenance actions on measurements. In difference to data warehouse systems, the users do data quality control and sensor calibration directly in the online system. The raw data is not changed but augmented by calibration equations and faulty data points are not deleted but marked as an error. Transformation functions convert direct observation to derived data, like discharge, on the fly. Improved stage-discharge relations apply directly to older measurements. The management system consists of a web portal, plotting and mapping facilities, import and export functions, an image database, and a management tool to assign tasks. A transparent link to CUAHSI Hydrological Information System (HIS), a data sharing portal, is currently under development using the standardized WaterML interface. The system is freely available and built upon open source tools. The system is in operational use for three observatories located in Germany, Ecuador and Kenya holding 10 to 50 Million records.
Gutiérrez, Manuel; Llobera, Andreu; Vila-Planas, Jordi; Capdevila, Fina; Demming, Stefanie; Büttgenbach, Stephanus; Mínguez, Santiago; Jiménez-Jorquera, Cecilia
2010-07-01
A multiparametric system able to classify red and white wines according to the grape varieties and for analysing some specific parameters is presented. The system, known as hybrid electronic tongue, consists of an array of electrochemical microsensors and a colorimetric optofluidic system. The array of electrochemical sensors is composed of six ISFETs based sensors, a conductivity sensor, a redox potential sensor and two amperometric electrodes, an Au microelectrode and a microelectrode for sensing electrochemical oxygen demand. The optofluidic system is entirely fabricated in polymer technology and comprises a hollow structure, air mirrors, microlenses and self-alignment structures. The data obtained from these sensors has been treated with multivariate advanced tools; Principal Component Analysis (PCA), for the patterning recognition and classification of wine samples, and Partial-Least Squares (PLS) regression, for quantification of several chemical and optical parameters of interest in wine quality. The results have demonstrated the utility of this system for distinguishing the samples according to the grape variety and year vintage and for quantifying several sample parameters of interest in wine quality control.
NASA Astrophysics Data System (ADS)
Ishizawa, Y.; Abe, K.; Shirako, G.; Takai, T.; Kato, H.
The electromagnetic compatibility (EMC) control method, system EMC analysis method, and system test method which have been applied to test the components of the MOS-1 satellite are described. The merits and demerits of the problem solving, specification, and system approaches to EMC control are summarized, and the data requirements of the SEMCAP (specification and electromagnetic compatibility analysis program) computer program for verifying the EMI safety margin of the components are sumamrized. Examples of EMC design are mentioned, and the EMC design process and selection method for EMC critical points are shown along with sample EMC test results.
Surface tension determination using liquid sample micromirror property
NASA Astrophysics Data System (ADS)
Hošek, Jan
2007-05-01
This paper presents an application of adaptive optics principle onto small sample of liquid surface tension measurement. The principle of experimental method devised by Ferguson (1924) is based on measurement of pressure difference across a liquid sample placed into small diameter capillary on condition of one flat meniscus of the liquid sample. Planarity or curvature radius of the capillary tip meniscus has to be measured and controlled, in order to fulfill this condition during measurement. Two different optical set-ups using liquid meniscus micromirror property are presented and its suitability for meniscus profile determination is compared. Meniscus radius optical measurement, data processing and control algorithm of the adaptive micromirror profile set are presented too. The presented adaptive optics system can be used for focal length control of microsystems based on liquid micromirrors or microlenses with long focal distances especially.
Investigating parameters participating in the infant respiratory control system attractor.
Terrill, Philip I; Wilson, Stephen J; Suresh, Sadasivam; Cooper, David M; Dakin, Carolyn
2008-01-01
Theoretically, any participating parameter in a non-linear system represents the dynamics of the whole system. Taken's time delay embedding theory provides the fundamental basis for allowing non-linear analysis to be performed on physiological, time-series data. In practice, only one measurable parameter is required to be measured to convey an accurate representation of the system dynamics. In this paper, the infant respiratory control system is represented using three variables-a digitally sampled respiratory inductive plethysmography waveform, and the derived parameters tidal volume and inter-breath interval time series data. For 14 healthy infants, these data streams were analysed using recurrence plot analysis across one night of sleep. The measured attractor size of these variables followed the same qualitative trends across the nights study. Results suggest that the attractor size measures of the derived IBI and tidal volume are representative surrogates for the raw respiratory waveform. The extent to which the relative attractor sizes of IBI and tidal volume remain constant through changing sleep state could potentially be used to quantify pathology, or maturation of breathing control.
NASA Astrophysics Data System (ADS)
Yussup, N.; Rahman, N. A. A.; Ibrahim, M. M.; Mokhtar, M.; Salim, N. A. A.; Soh@Shaari, S. C.; Azman, A.
2017-01-01
Neutron Activation Analysis (NAA) process has been established in Malaysian Nuclear Agency (Nuclear Malaysia) since 1980s. Most of the procedures established especially from sample registration to sample analysis are performed manually. These manual procedures carried out by the NAA laboratory personnel are time consuming and inefficient. Hence, a software to support the system automation is developed to provide an effective method to replace redundant manual data entries and produce faster sample analysis and calculation process. This paper describes the design and development of automation software for NAA process which consists of three sub-programs. The sub-programs are sample registration, hardware control and data acquisition; and sample analysis. The data flow and connection between the sub-programs will be explained. The software is developed by using National Instrument LabView development package.
Applying Online Monitoring for Nuclear Power Plant Instrumentation and Control
NASA Astrophysics Data System (ADS)
Hashemian, H. M.
2010-10-01
This paper presents a practical review of the state-of-the-art means for applying OLM data acquisition in nuclear power plant instrumentation and control, qualifying or validating the OLM data, and then analyzing it for static and dynamic performance monitoring applications. Whereas data acquisition for static or steady-state OLM applications can require sample rates of anywhere from 1 to 10 seconds to 1 minutes per sample, for dynamic data acquisition, higher sampling frequencies are required (e.g., 100 to 1000 Hz) using a dedicated data acquisition system capable of providing isolation, anti-aliasing and removal of extraneous noise, and analog-to-digital (A/D) conversion. Qualifying the data for use with OLM algorithms can involve removing data `dead' spots (for static data) and calculating, examining, and trending amplitude probability density, variance, skewness, and kurtosis. For static OLM applications with redundant signals, trending and averaging qualification techniques are used, and for single or non-redundant signals physical and empirical modeling are used. Dynamic OLM analysis is performed in the frequency domain and/or time domain, and is based on the assumption that sensors' or transmitters' dynamic characteristics are linear and that the input noise signal (i.e., the process fluctuations) has proper spectral characteristics.
Level 1 environmental assessment performance evaluation. Final report jun 77-oct 78
DOE Office of Scientific and Technical Information (OSTI.GOV)
Estes, E.D.; Smith, F.; Wagoner, D.E.
1979-02-01
The report gives results of a two-phased evaluation of Level 1 environmental assessment procedures. Results from Phase I, a field evaluation of the Source Assessment Sampling System (SASS), showed that the SASS train performed well within the desired factor of 3 Level 1 accuracy limit. Three sample runs were made with two SASS trains sampling simultaneously and from approximately the same sampling point in a horizontal duct. A Method-5 train was used to estimate the 'true' particulate loading. The sampling systems were upstream of the control devices to ensure collection of sufficient material for comparison of total particulate, particle sizemore » distribution, organic classes, and trace elements. Phase II consisted of providing each of three organizations with three types of control samples to challenge the spectrum of Level 1 analytical procedures: an artificial sample in methylene chloride, an artificial sample on a flyash matrix, and a real sample composed of the combined XAD-2 resin extracts from all Phase I runs. Phase II results showed that when the Level 1 analytical procedures are carefully applied, data of acceptable accuracy is obtained. Estimates of intralaboratory and interlaboratory precision are made.« less
NASA Astrophysics Data System (ADS)
Noyes, Ben F.; Mokaberi, Babak; Mandoy, Ram; Pate, Alex; Huijgen, Ralph; McBurney, Mike; Chen, Owen
2017-03-01
Reducing overlay error via an accurate APC feedback system is one of the main challenges in high volume production of the current and future nodes in the semiconductor industry. The overlay feedback system directly affects the number of dies meeting overlay specification and the number of layers requiring dedicated exposure tools through the fabrication flow. Increasing the former number and reducing the latter number is beneficial for the overall efficiency and yield of the fabrication process. An overlay feedback system requires accurate determination of the overlay error, or fingerprint, on exposed wafers in order to determine corrections to be automatically and dynamically applied to the exposure of future wafers. Since current and future nodes require correction per exposure (CPE), the resolution of the overlay fingerprint must be high enough to accommodate CPE in the overlay feedback system, or overlay control module (OCM). Determining a high resolution fingerprint from measured data requires extremely dense overlay sampling that takes a significant amount of measurement time. For static corrections this is acceptable, but in an automated dynamic correction system this method creates extreme bottlenecks for the throughput of said system as new lots have to wait until the previous lot is measured. One solution is using a less dense overlay sampling scheme and employing computationally up-sampled data to a dense fingerprint. That method uses a global fingerprint model over the entire wafer; measured localized overlay errors are therefore not always represented in its up-sampled output. This paper will discuss a hybrid system shown in Fig. 1 that combines a computationally up-sampled fingerprint with the measured data to more accurately capture the actual fingerprint, including local overlay errors. Such a hybrid system is shown to result in reduced modelled residuals while determining the fingerprint, and better on-product overlay performance.
Controlling Processes on Carbonate Chemistry across the Pacific
NASA Astrophysics Data System (ADS)
Hartman, S. E.
2016-12-01
The SWIRE NOC Ocean Monitoring System (SNOMS) project is an innovative programme helping to answer important questions about global climate change by using a commercial ship of opportunity to measure carbon in the surface of the ocean. Daily sampling coupled to continuous underway observation from a ship of opportunity (MV Shengking) provides new insights into the processes controlling variability in the carbonate system across the Pacific. The ships track runs from Vancouver (Canada) to Brisbane (Australia). Daily samples were taken on-board and measurements of Total alkalinity (TA) and total dissolved inorganic carbon (DIC) were determined. This was alongside measurements of nutrients and continuous records of temperature, salinity, chlorophyll-fluorescence, carbon dioxide and dissolved oxygen (DO). These sensor based measurements were validated using the discrete samples. Carbon dioxide calculated from DIC and TA showed an offset from the sensor data of up to 8uatm. This and comparisons with climatology were used to calibrate the sensor data. The data have been compared with previous data from the MV Pacific Celebes that ran a similar route until 2012. The data show a clear increase in seawater carbon dioxide, tracking the atmospheric increases. Along track the partial pressure of seawater carbon dioxide varied by over 150 uatm. The highest values were seen just south of the equator in the Pacific, which is an important source region for carbon dioxide to the atmosphere.
Chen, Hsin-Yung; Wu, Jin-Shang; Hyland, Brian; Lu, Xiao-Dong; Chen, Jia Jin Jason
2008-08-01
The use of cables for recording neural activity limits the scope of behavioral tests used in conscious free-moving animals. Particularly, cable attachments make it impossible to record in three-dimensional (3D) mazes where levels are vertically stacked or in enclosed spaces. Such environments are of particular interest in investigations of hippocampal place cells, in which neural activity is correlated with spatial position in the environment. We developed a flexible miniaturized Bluetooth-based wireless data acquisition system. The wireless module included an 8-channel analogue front end, digital controller, and Bluetooth transceiver mounted on a backpack. Our bidirectional wireless design allowed all data channels to be previewed at 1 kHz sample rate, and one channel, selected by remote control, to be sampled at 10 kHz. Extracellular recordings of neuronal activity are highly susceptible to ambient electrical noise due to the high electrode impedance. Through careful design of appropriate shielding and hardware configuration to avoid ground loops, mains power and Bluetooth hopping frequency noise were reduced sufficiently to yield signal quality comparable to those recorded by wired systems. With this system we were able to obtain single-unit recordings of hippocampal place cells in rats running an enclosed vertical maze, over a range of 5 m.
Routine sampling and the control of Legionella spp. in cooling tower water systems.
Bentham, R H
2000-10-01
Cooling water samples from 31 cooling tower systems were cultured for Legionella over a 16-week summer period. The selected systems were known to be colonized by Legionella. Mean Legionella counts and standard deviations were calculated and time series correlograms prepared for each system. The standard deviations of Legionella counts in all the systems were very large, indicating great variability in the systems over the time period. Time series analyses demonstrated that in the majority of cases there was no significant relationship between the Legionella counts in the cooling tower at time of collection and the culture result once it was available. In the majority of systems (25/28), culture results from Legionella samples taken from the same systems 2 weeks apart were not statistically related. The data suggest that determinations of health risks from cooling towers cannot be reliably based upon single or infrequent Legionella tests.
NASA Astrophysics Data System (ADS)
Soni, Jigensh; Yadav, R. K.; Patel, A.; Gahlaut, A.; Mistry, H.; Parmar, K. G.; Mahesh, V.; Parmar, D.; Prajapati, B.; Singh, M. J.; Bandyopadhyay, M.; Bansal, G.; Pandya, K.; Chakraborty, A.
2013-02-01
Twin Source - An Inductively coupled two RF driver based 180 kW, 1 MHz negative ion source experimental setup is initiated at IPR, Gandhinagar, under Indian program, with the objective of understanding the physics and technology of multi-driver coupling. Twin Source [1] (TS) also provides an intermediate platform between operational ROBIN [2] [5] and eight RF drivers based Indian test facility -INTF [3]. A twin source experiment requires a central system to provide control, data acquisition and communication interface, referred as TS-CODAC, for which a software architecture similar to ITER CODAC core system has been decided for implementation. The Core System is a software suite for ITER plant system manufacturers to use as a template for the development of their interface with CODAC. The ITER approach, in terms of technology, has been adopted for the TS-CODAC so as to develop necessary expertise for developing and operating a control system based on the ITER guidelines as similar configuration needs to be implemented for the INTF. This cost effective approach will provide an opportunity to evaluate and learn ITER CODAC technology, documentation, information technology and control system processes, on an operational machine. Conceptual design of the TS-CODAC system has been completed. For complete control of the system, approximately 200 Nos. control signals and 152 acquisition signals are needed. In TS-CODAC, control loop time required is within the range of 5ms - 10 ms, therefore for the control system, PLC (Siemens S-7 400) has been chosen as suggested in the ITER slow controller catalog. For the data acquisition, the maximum sampling interval required is 100 micro second, and therefore National Instruments (NI) PXIe system and NI 6259 digitizer cards have been selected as suggested in the ITER fast controller catalog. This paper will present conceptual design of TS -CODAC system based on ITER CODAC Core software and applicable plant system integration processes.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Castaldini, C.; DeRosier, R.; Waterland, L.R.
1986-02-01
The report discusses emission results from comprehensive flue-gas sampling of a gas- and oil-fired industrial boiler equipped with Exxon's Thermal DeNOx Ammonia Injection Process for NOx reduction. Comprehensive emission measurements included continuous monitoring of flue gas emissions; source assessment sampling system (SASS) tests; EPA Method 5/17 for solid and condensible particulate emissions and ammonia emissions; controlled condensation system for SO2 and SO3; and N2O emission sampling. Ammonia injection at a NH3/NO molar ratio of 2.52 gave a NOx reduction of 41% from an uncontrolled level of 234 ppm to a controlled level of 137 ppm. NH3 emissions increased from 11more » ppm for the baseline to an average of 430 ppm for ammonia injection. Nitrous oxide, N2O, was reduced 68% from a 50 ppm baseline level to a 17 ppm controlled level. Total particulate emissions increased by an order of magnitude from a baseline of 17.7 ng/J to a controlled level of 182 ng/J.« less
A system architecture for a planetary rover
NASA Technical Reports Server (NTRS)
Smith, D. B.; Matijevic, J. R.
1989-01-01
Each planetary mission requires a complex space vehicle which integrates several functions to accomplish the mission and science objectives. A Mars Rover is one of these vehicles, and extends the normal spacecraft functionality with two additional functions: surface mobility and sample acquisition. All functions are assembled into a hierarchical and structured format to understand the complexities of interactions between functions during different mission times. It can graphically show data flow between functions, and most importantly, the necessary control flow to avoid unambiguous results. Diagrams are presented organizing the functions into a structured, block format where each block represents a major function at the system level. As such, there are six blocks representing telecomm, power, thermal, science, mobility and sampling under a supervisory block called Data Management/Executive. Each block is a simple collection of state machines arranged into a hierarchical order very close to the NASREM model for Telerobotics. Each layer within a block represents a level of control for a set of state machines that do the three primary interface functions: command, telemetry, and fault protection. This latter function is expanded to include automatic reactions to the environment as well as internal faults. Lastly, diagrams are presented that trace the system operations involved in moving from site to site after site selection. The diagrams clearly illustrate both the data and control flows. They also illustrate inter-block data transfers and a hierarchical approach to fault protection. This systems architecture can be used to determine functional requirements, interface specifications and be used as a mechanism for grouping subsystems (i.e., collecting groups of machines, or blocks consistent with good and testable implementations).
Monitoring of beer fermentation based on hybrid electronic tongue.
Kutyła-Olesiuk, Anna; Zaborowski, Michał; Prokaryn, Piotr; Ciosek, Patrycja
2012-10-01
Monitoring of biotechnological processes, including fermentation is extremely important because of the rapidly occurring changes in the composition of the samples during the production. In the case of beer, the analysis of physicochemical parameters allows for the determination of the stage of fermentation process and the control of its possible perturbations. As a tool to control the beer production process a sensor array can be used, composed of potentiometric and voltammetric sensors (so-called hybrid Electronic Tongue, h-ET). The aim of this study is to apply electronic tongue system to distinguish samples obtained during alcoholic fermentation. The samples originate from batch of homemade beer fermentation and from two stages of the process: fermentation reaction and maturation of beer. The applied sensor array consists of 10 miniaturized ion-selective electrodes (potentiometric ET) and silicon based 3-electrode voltammetric transducers (voltammetric ET). The obtained results were processed using Partial Least Squares (PLS) and Partial Least Squares-Discriminant Analysis (PLS-DA). For potentiometric data, voltammetric data, and combined potentiometric and voltammetric data, comparison of the classification ability was conducted based on Root Mean Squared Error (RMSE), sensitivity, specificity, and coefficient F calculation. It is shown, that in the contrast to the separately used techniques, the developed hybrid system allowed for a better characterization of the beer samples. Data fusion in hybrid ET enables to obtain better results both in qualitative analysis (RMSE, specificity, sensitivity) and in quantitative analysis (RMSE, R(2), a, b). Copyright © 2012 Elsevier B.V. All rights reserved.
Ultrasonic scanning system for imaging flaw growth in composites
NASA Technical Reports Server (NTRS)
Kiraly, L. J.; Meyn, E. H.
1982-01-01
A system for measuring and visually representing damage in composite specimens while they are being loaded was demonstrated. It uses a hobbiest grade microcomputer system to control data taking and image processing. The system scans operator selected regions of the specimen while it is under load in a tensile test machine and measures internal damage by the attenuation of a 2.5 MHz ultrasonic beam passed through the specimen. The microcomputer dynamically controls the position of ultrasonic transducers mounted on a two axis motor driven carriage. As many as 65,536 samples can be taken and filed on a floppy disk system in less than four minutes.
Sampling and Visualizing Creases with Scale-Space Particles
Kindlmann, Gordon L.; Estépar, Raúl San José; Smith, Stephen M.; Westin, Carl-Fredrik
2010-01-01
Particle systems have gained importance as a methodology for sampling implicit surfaces and segmented objects to improve mesh generation and shape analysis. We propose that particle systems have a significantly more general role in sampling structure from unsegmented data. We describe a particle system that computes samplings of crease features (i.e. ridges and valleys, as lines or surfaces) that effectively represent many anatomical structures in scanned medical data. Because structure naturally exists at a range of sizes relative to the image resolution, computer vision has developed the theory of scale-space, which considers an n-D image as an (n + 1)-D stack of images at different blurring levels. Our scale-space particles move through continuous four-dimensional scale-space according to spatial constraints imposed by the crease features, a particle-image energy that draws particles towards scales of maximal feature strength, and an inter-particle energy that controls sampling density in space and scale. To make scale-space practical for large three-dimensional data, we present a spline-based interpolation across scale from a small number of pre-computed blurrings at optimally selected scales. The configuration of the particle system is visualized with tensor glyphs that display information about the local Hessian of the image, and the scale of the particle. We use scale-space particles to sample the complex three-dimensional branching structure of airways in lung CT, and the major white matter structures in brain DTI. PMID:19834216
Lefkowitz, Daniel; Pechter, Elise; Fitzsimmons, Kathleen; Lumia, Margaret; Stephens, Alicia C; Davis, Letitia; Flattery, Jennifer; Weinberg, Justine; Harrison, Robert J; Reilly, Mary Jo; Filios, Margaret S; White, Gretchen E; Rosenman, Kenneth D
2015-11-01
Isocyanates remain a leading cause of work-related asthma (WRA). Two independent data systems were analyzed for the period 1993-2008: (1) State-based WRA case surveillance data on persons with isocyanate-induced WRA from four states, and (2) Occupational Safety and Health Administration (OSHA) Integrated Management Information System (IMIS) isocyanate air sampling results. We identified 368 cases of isocyanate-induced WRA from 32 industries and 678 OSHA isocyanate air samples with detectable levels from 31 industries. Seventeen industries were unique to one or the other dataset. Isocyanate-induced WRA continues to occur in a wide variety of industries. Two data systems uncovered industries with isocyanate exposures and/or illness. Improved control measures and standards, including medical surveillance, are needed. More emphasis is needed on task-specific guidance, spill clean-up procedures, skin and respiratory protection, and targeted medical monitoring to mitigate the hazards of isocyanate use. © 2015 Wiley Periodicals, Inc.
Dynamics and Robust Control of Sampled Data Systems for Large Space Structures
1992-11-01
physical interpretation of J 1 is this: We wish to keep the state near zero without excessive control-energy expenditcure. The weighting matrix, Q...can be given as follows. Defining v(k) -- [RI+ HT75(k÷!) H) -’ HTP (k+l)Gx(k) (249) where P(k) is a modified version of the Ricatti Equation. £(k)-C... Manual ", GTICES Systems Laboratory, Georgia Institute of Technology, Altanta, GA, Rev.J, April 1978. 21) Ericsson, A.J., "Determination of Frequencies
Aircraft adaptive learning control
NASA Technical Reports Server (NTRS)
Lee, P. S. T.; Vanlandingham, H. F.
1979-01-01
The optimal control theory of stochastic linear systems is discussed in terms of the advantages of distributed-control systems, and the control of randomly-sampled systems. An optimal solution to longitudinal control is derived and applied to the F-8 DFBW aircraft. A randomly-sampled linear process model with additive process and noise is developed.
An optical motion measuring system for laterally oscillated fatigue tests
NASA Technical Reports Server (NTRS)
Tripp, John S.; Tcheng, Ping; Murri, Gretchen B.; Sharpe, Scott
1993-01-01
This paper describes an optical system developed for materials testing laboratories at NASA Langley Research Center (LaRC) for high resolution monitoring of the transverse displacement and angular rotation of a test specimen installed in an axial-tension bending machine (ATB) during fatigue tests. It consists of a small laser, optics, a motorized mirror, three photodiodes, electronic detection and counting circuits, a data acquisition system, and a personal computer. A 3-inch by 5-inch rectangular plate attached to the upper grip of the test machine serves as a target base for the optical system. The personal computer automates the fatigue test procedure, controls data acquisition, performs data reduction, and provides user displays. The data acquisition system also monitors signals from up to 16 strain gages mounted on the test specimen. The motion measuring system is designed to continuously monitor and correlate the amplitude of the oscillatory motion with the strain gage signals in order to detect the onset of failure of the composite test specimen. A prototype system has been developed and tested which exceeds the design specifications of +/- 0.01 inch displacement accuracy, and +/- 0.25 deg angular accuracy at a sampling rate of 100 samples per second.
Riddick, L; Simbanin, C
2001-01-01
EPA is conducting a National Study of Chemical Residues in Lake Fish Tissue. The study involves five analytical laboratories, multiple sampling teams from each of the 47 participating states, several tribes, all 10 EPA Regions and several EPA program offices, with input from other federal agencies. To fulfill study objectives, state and tribal sampling teams are voluntarily collecting predator and bottom-dwelling fish from approximately 500 randomly selected lakes over a 4-year period. The fish will be analyzed for more than 300 pollutants. The long-term nature of the study, combined with the large number of participants, created several QA challenges: (1) controlling variability among sampling activities performed by different sampling teams from more than 50 organizations over a 4-year period; (2) controlling variability in lab processes over a 4-year period; (3) generating results that will meet the primary study objectives for use by OW statisticians; (4) generating results that will meet the undefined needs of more than 50 participating organizations; and (5) devising a system for evaluating and defining data quality and for reporting data quality assessments concurrently with the data to ensure that assessment efforts are streamlined and that assessments are consistent among organizations. This paper describes the QA program employed for the study and presents an interim assessment of the program's effectiveness.
Contained radiological analytical chemistry module
Barney, David M.
1989-01-01
A system which provides analytical determination of a plurality of water chemistry parameters with respect to water samples subject to radiological contamination. The system includes a water sample analyzer disposed within a containment and comprising a sampling section for providing predetermined volumes of samples for analysis; a flow control section for controlling the flow through the system; and a gas analysis section for analyzing samples provided by the sampling system. The sampling section includes a controllable multiple port valve for, in one position, metering out sample of a predetermined volume and for, in a second position, delivering the material sample for analysis. The flow control section includes a regulator valve for reducing the pressure in a portion of the system to provide a low pressure region, and measurement devices located in the low pressure region for measuring sample parameters such as pH and conductivity, at low pressure. The gas analysis section which is of independent utility provides for isolating a small water sample and extracting the dissolved gases therefrom into a small expansion volume wherein the gas pressure and thermoconductivity of the extracted gas are measured.
Contained radiological analytical chemistry module
Barney, David M.
1990-01-01
A system which provides analytical determination of a plurality of water chemistry parameters with respect to water samples subject to radiological contamination. The system includes a water sample analyzer disposed within a containment and comprising a sampling section for providing predetermined volumes of samples for analysis; a flow control section for controlling the flow through the system; and a gas analysis section for analyzing samples provided by the sampling system. The sampling section includes a controllable multiple port valve for, in one position, metering out sample of a predetermined volume and for, in a second position, delivering the material sample for analysis. The flow control section includes a regulator valve for reducing the pressure in a portion of the system to provide a low pressure region, and measurement devices located in the low pressure region for measuring sample parameters such as pH and conductivity, at low pressure. The gas analysis section which is of independent utility provides for isolating a small water sample and extracting the dissolved gases therefrom into a small expansion volume wherein the gas pressure and thermoconductivity of the extracted gas are measured.
Systematic on-site monitoring of compliance dust samples
DOE Office of Scientific and Technical Information (OSTI.GOV)
Grayson, R.L.; Gandy, J.R.
1996-12-31
Maintaining compliance with U.S. respirable coal mine dust standards can be difficult on high-productivity longwall panels. Comprehensive and systematic analysis of compliance dust sample data, coupled with access to the U.S. Bureau of Mines (USBM) DUSTPRO, can yield important information for use in maintaining compliance. The objective of this study was to develop and apply a customized software for the collection, storage, modification, and analysis of respirable dust data while providing for flexible export of data and linking with the USBM`s expert advisory system on dust control. An executable, IBM-compatible software was created and customized for use by the personmore » in charge of collecting, submitting, analyzing, and monitoring respirable dust compliance samples. Both descriptive statistics and multiple regression analysis were incorporated. The software allows ASCH files to be exported and directly links with DUSTPRO. After development and validation of the software, longwall compliance data from two different mines was analyzed to evaluate the value of the software. Data included variables on respirable dust concentration, tons produced, the existence of roof/floor rock (dummy variable), and the sampling cycle (dummy variables). Because of confidentiality, specific data will not be presented, only the equations and ANOVA tables. The final regression models explained 83.8% and 61.1% of the variation in the data for the two panels. Important correlations among variables within sampling cycles showed the value of using dummy variables for sampling cycles. The software proved flexible and fast for its intended use. The insights obtained from use improved the systematic monitoring of respirable dust compliance data, especially for pinpointing the most effective dust control methods during specific sampling cycles.« less
NASA aviation safety reporting system
NASA Technical Reports Server (NTRS)
1977-01-01
During the third quarter of operation of the Aviation Safety Reporting System (ASRS), 1429 reports concerning aviation safety were received from pilots, air traffic controllers, and others in the national aviation system. Details of the administration and results of the program are discussed. The design and construction of the ASRS data base are briefly presented. Altitude deviations and potential aircraft conflicts associated with misunderstood clearances were studied and the results are discussed. Summary data regarding alert bulletins, examples of alert bulletins and responses to them, and a sample of deidentified ASRS reports are provided.
Electronic readout system for the Belle II imaging Time-Of-Propagation detector
NASA Astrophysics Data System (ADS)
Kotchetkov, Dmitri
2017-07-01
The imaging Time-Of-Propagation (iTOP) detector, constructed for the Belle II experiment at the SuperKEKB e+e- collider, is an 8192-channel high precision Cherenkov particle identification detector with timing resolution below 50 ps. To acquire data from the iTOP, a novel front-end electronic readout system was designed, built, and integrated. Switched-capacitor array application-specific integrated circuits are used to sample analog signals. Triggering, digitization, readout, and data transfer are controlled by Xilinx Zynq-7000 system on a chip devices.
Aircraft attitude measurement using a vector magnetometer
NASA Technical Reports Server (NTRS)
Peitila, R.; Dunn, W. R., Jr.
1977-01-01
The feasibility of a vector magnetometer system was investigated by developing a technique to determine attitude given magnetic field components. Sample calculations are then made using the earth's magnetic field data acquired during actual flight conditions. Results of these calculations are compared graphically with measured attitude data acquired simultaneously with the magnetic data. The role and possible implementation of various reference angles are discussed along with other pertinent considerations. Finally, it is concluded that the earth's magnetic field as measured by modern vector magnetometers can play a significant role in attitude control systems.
Mobile phone based SCADA for industrial automation.
Ozdemir, Engin; Karacor, Mevlut
2006-01-01
SCADA is the acronym for "Supervisory Control And Data Acquisition." SCADA systems are widely used in industry for supervisory control and data acquisition of industrial processes. Conventional SCADA systems use PC, notebook, thin client, and PDA as a client. In this paper, a Java-enabled mobile phone has been used as a client in a sample SCADA application in order to display and supervise the position of a sample prototype crane. The paper presents an actual implementation of the on-line controlling of the prototype crane via mobile phone. The wireless communication between the mobile phone and the SCADA server is performed by means of a base station via general packet radio service (GPRS) and wireless application protocol (WAP). Test results have indicated that the mobile phone based SCADA integration using the GPRS or WAP transfer scheme could enhance the performance of the crane in a day without causing an increase in the response times of SCADA functions. The operator can visualize and modify the plant parameters using his mobile phone, without reaching the site. In this way maintenance costs are reduced and productivity is increased.
On the hitchhiker Robot Operated Materials Processing System: Experiment data system
NASA Technical Reports Server (NTRS)
Kizhner, Semion; Jenstrom, Del
1995-01-01
The Space Shuttle Discovery STS-64 mission carried the first American autonomous robot into space, the Robot Operated Materials Processing System (ROMPS). On this mission ROMPS was the only Hitchhiker experiment and had a unique opportunity to utilize all Hitchhiker space carrier capabilities. ROMPS conducted rapid thermal processing of the one hundred semiconductor material samples to study how micro gravity affects the resulting material properties. The experiment was designed, built and operated by a small GSFC team in cooperation with industry and university based principal investigators who provided the material samples and data interpretation. ROMPS' success presents some valuable lessons in such cooperation, as well as in the utilization of the Hitchhiker carrier for complex applications. The motivation of this paper is to share these lessons with the scientific community interested in attached payload experiments. ROMPS has a versatile and intelligent material processing control data system. This paper uses the ROMPS data system as the guiding thread to present the ROMPS mission experience. It presents an overview of the ROMPS experiment followed by considerations of the flight and ground data subsystems and their architecture, data products generation during mission operations, and post mission data utilization. It then presents the lessons learned from the development and operation of the ROMPS data system as well as those learned during post-flight data processing.
New multirate sampled-data control law structure and synthesis algorithm
NASA Technical Reports Server (NTRS)
Berg, Martin C.; Mason, Gregory S.; Yang, Gen-Sheng
1992-01-01
A new multirate sampled-data control law structure is defined and a new parameter-optimization-based synthesis algorithm for that structure is introduced. The synthesis algorithm can be applied to multirate, multiple-input/multiple-output, sampled-data control laws having a prescribed dynamic order and structure, and a priori specified sampling/update rates for all sensors, processor states, and control inputs. The synthesis algorithm is applied to design two-input, two-output tip position controllers of various dynamic orders for a sixth-order, two-link robot arm model.
Information management in Iranian Maternal Mortality Surveillance System.
Sadoughi, Farahnaz; Karimi, Afsaneh; Erfannia, Leila
2017-07-01
Maternal mortality is preventable by proper information management and is the main target of the Maternal Mortality Surveillance System (MMSS). This study aimed to determine the status of information management in the Iranian Maternal Mortality Surveillance System (IMMSS). The population of this descriptive and analytical study, which was conducted in 2016, included 96 administrative staff of health and treatment deputies of universities of medical sciences and the Ministry of Health in Iran. Data were gathered by a five-part questionnaire with confirmed validity and reliability. A total of 76 questionnaires were completed, and data were analyzed using SPSS software, version 19, by descriptive and inferential statistics. The relationship between variables "organizational unit" and the four studied axes was studied using Kendall's correlation coefficient test. The status of information management in IMMSS was desirable. Data gathering and storage axis and data processing and compilation axis achieved the highest (2.7±0.46) and the lowest (2.4±0.49) mean scores, respectively. The data-gathering method, control of a sample of women deaths in reproductive age in the universities of medical sciences, use of international classification of disease, and use of this system information by management teams to set resources allocation achieved the lowest mean scores in studied axes. Treatment deputy staff had a more positive attitude toward the status of information management of IMMSS than the health deputy staff (p=0.004). Although the status of information management in IMMSS was desirable, it could be improved by modification of the data-gathering method; creating communication links between different data resources; a periodic sample control of women deaths in reproductive age in the universities of medical sciences; and implementing ICD-MM and integration of its rules on a unified system of death.
Modeling the uncertainty of estimating forest carbon stocks in China
NASA Astrophysics Data System (ADS)
Yue, T. X.; Wang, Y. F.; Du, Z. P.; Zhao, M. W.; Zhang, L. L.; Zhao, N.; Lu, M.; Larocque, G. R.; Wilson, J. P.
2015-12-01
Earth surface systems are controlled by a combination of global and local factors, which cannot be understood without accounting for both the local and global components. The system dynamics cannot be recovered from the global or local controls alone. Ground forest inventory is able to accurately estimate forest carbon stocks at sample plots, but these sample plots are too sparse to support the spatial simulation of carbon stocks with required accuracy. Satellite observation is an important source of global information for the simulation of carbon stocks. Satellite remote-sensing can supply spatially continuous information about the surface of forest carbon stocks, which is impossible from ground-based investigations, but their description has considerable uncertainty. In this paper, we validated the Lund-Potsdam-Jena dynamic global vegetation model (LPJ), the Kriging method for spatial interpolation of ground sample plots and a satellite-observation-based approach as well as an approach for fusing the ground sample plots with satellite observations and an assimilation method for incorporating the ground sample plots into LPJ. The validation results indicated that both the data fusion and data assimilation approaches reduced the uncertainty of estimating carbon stocks. The data fusion had the lowest uncertainty by using an existing method for high accuracy surface modeling to fuse the ground sample plots with the satellite observations (HASM-SOA). The estimates produced with HASM-SOA were 26.1 and 28.4 % more accurate than the satellite-based approach and spatial interpolation of the sample plots, respectively. Forest carbon stocks of 7.08 Pg were estimated for China during the period from 2004 to 2008, an increase of 2.24 Pg from 1984 to 2008, using the preferred HASM-SOA method.
Operational Processing of Ground Validation Data for the Tropical Rainfall Measuring Mission
NASA Technical Reports Server (NTRS)
Kulie, Mark S.; Robinson, Mike; Marks, David A.; Ferrier, Brad S.; Rosenfeld, Danny; Wolff, David B.
1999-01-01
The Tropical Rainfall Measuring Mission (TRMM) satellite was successfully launched in November 1997. A primary goal of TRMM is to sample tropical rainfall using the first active spaceborne precipitation radar. To validate TRMM satellite observations, a comprehensive Ground Validation (GV) Program has been implemented for this mission. A key component of GV is the analysis and quality control of meteorological ground-based radar data from four primary sites: Melbourne, FL; Houston, TX; Darwin, Australia; and Kwajalein Atoll, RMI. As part of the TRMM GV effort, the Joint Center for Earth Systems Technology (JCET) at the University of Maryland, Baltimore County, has been tasked with developing and implementing an operational system to quality control (QC), archive, and provide data for subsequent rainfall product generation from the four primary GV sites. This paper provides an overview of the JCET operational environment. A description of the QC algorithm and performance, in addition to the data flow procedure between JCET and the TRNM science and Data Information System (TSDIS), are presented. The impact of quality-controlled data on higher level rainfall and reflectivity products will also be addressed, Finally, a brief description of JCET's expanded role into producing reference rainfall products will be discussed.
Automated Drug Identification for Urban Hospitals
NASA Technical Reports Server (NTRS)
Shirley, Donna L.
1971-01-01
Many urban hospitals are becoming overloaded with drug abuse cases requiring chemical analysis for identification of drugs. In this paper, the requirements for chemical analysis of body fluids for drugs are determined and a system model for automated drug analysis is selected. The system as modeled, would perform chemical preparation of samples, gas-liquid chromatographic separation of drugs in the chemically prepared samples, infrared spectrophotometric analysis of the drugs, and would utilize automatic data processing and control for drug identification. Requirements of cost, maintainability, reliability, flexibility, and operability are considered.
[Information technology for the management of health care data: the EPIweb project].
Vittorini, Pierpaolo; Necozione, Stefano; di Orio, Ferdinando
2005-01-01
In the US, the Center for Disease Control and Prevention has produced has increased the permeability of the computer science technologies, in order to achieve a better and more efficient management of health care data. In this context, the present paper proposes a discussion regarding a web-based information system, called EPIweb. This system allows researchers to select the centers for the data entry, collect and elaborate health care data, produce technical reports and discuss results. Such a system aims to be easy-to-use, totally configurable and particularly suitable for the management of multicenter studies. The paper shows the EPIweb features, proposes a sample system run, and concludes with a discussion regarding both the advantages and the possible improvements and extensions.
Current Status of the Beam Position Monitoring System at TLS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kuo, C. H.; Hu, K. H.; Chen, Jenny
2006-11-20
The beam position monitoring system is an important part of a synchrotron light source that supports its routine operation and studies of beam physics. The Taiwan light source is equipped with 59 BPMs. Highly precise closed orbits are measured by multiplexing BPMs. Data are acquired using multi-channel 16-bit ADC modules. Orbit data are sampled every millisecond. Fast orbit data are shared in a reflective memory network to support fast orbit feedback. Averaged data were updated to control database at a rate of 10 Hz. A few new generation digital BPMs were tested to evaluate their performance and functionality. This reportmore » summarizes the system structure, the software environment and the preliminary beam test of the BPM system.« less
Current Status of the Beam Position Monitoring System at TLS
NASA Astrophysics Data System (ADS)
Kuo, C. H.; Hu, K. H.; Chen, Jenny; Lee, Demi; Wang, C. J.; Hsu, S. Y.; Hsu, K. T.
2006-11-01
The beam position monitoring system is an important part of a synchrotron light source that supports its routine operation and studies of beam physics. The Taiwan light source is equipped with 59 BPMs. Highly precise closed orbits are measured by multiplexing BPMs. Data are acquired using multi-channel 16-bit ADC modules. Orbit data are sampled every millisecond. Fast orbit data are shared in a reflective memory network to support fast orbit feedback. Averaged data were updated to control database at a rate of 10 Hz. A few new generation digital BPMs were tested to evaluate their performance and functionality. This report summarizes the system structure, the software environment and the preliminary beam test of the BPM system.
Controller and data acquisition system for SIDECAR ASIC driven HAWAII detectors
NASA Astrophysics Data System (ADS)
Ramaprakash, Anamparambu; Burse, Mahesh; Chordia, Pravin; Chillal, Kalpesh; Kohok, Abhay; Mestry, Vilas; Punnadi, Sujit; Sinha, Sakya
2010-07-01
SIDECAR is an Application Specific Integrated Circuit (ASIC), which can be used for control and data acquisition from near-IR HAWAII detectors offered by Teledyne Imaging Sensors (TIS), USA. The standard interfaces provided by Teledyne are COM API and socket servers running under MS Windows platform. These interfaces communicate to the ASIC (and the detector) through an intermediate card called JWST ASIC Drive Electronics (JADE2). As part of an ongoing programme of several years, for developing astronomical focal plane array (CCDs, CMOS and Hybrid) controllers and data acquisition systems (CDAQs), IUCAA is currently developing the next generation controllers employing Virtex-5 family FPGA devices. We present here the capabilities which are built into these new CDAQs for handling HAWAII detectors. In our system, the computer which hosts the application programme, user interface and device drivers runs on a Linux platform. It communicates through a hot-pluggable USB interface (with an optional optical fibre extender) to the FPGA-based card which replaces the JADE2. The FPGA board in turn, controls the SIDECAR ASIC and through it a HAWAII-2RG detector, both of which are located in a cryogenic test Dewar set up which is liquid nitrogen cooled. The system can acquire data over 1, 4, or 32 readout channels, with or without binning, at different speeds, can define sub-regions for readout, offers various readout schemes like Fowler sampling, up-theramp etc. In this paper, we present the performance results obtained from a prototype system.
MSFC Electrostatic Levitator (ESL) Rapid Quench System
NASA Technical Reports Server (NTRS)
SanSoucie, Michael P.; Craven, Paul D.; Rogers, Jan R.
2014-01-01
The NASA Marshall Space Flight Center (MSFC) Electrostatic Levitator (ESL) Laboratory is a unique facility for investigators studying high-temperature materials. The laboratory boasts two levitators in which samples can be levitated, heated, melted, undercooled, and resolidified, all without the interference of a container or data-gathering instrument. The ESL main chamber has been upgraded with the addition of a rapid quench system. This system allows samples to be dropped into a quench vessel that can be filled with a low melting point material, such as a gallium or indium alloy. Thereby allowing rapid quenching of undercooled liquid metals. Up to 8 quench vessels can be loaded into the quench wheel, which is indexed with LabVIEW control software. This allows up to 8 samples to be rapidly quenched before having to open the chamber. The system has been tested successfully on several zirconium samples. Future work will be done with other materials using different quench mediums. Microstructural analysis will also be done on successfully quench samples.
Formation Flying Control Implementation in Highly Elliptical Orbits
NASA Technical Reports Server (NTRS)
Capo-Lugo, Pedro A.; Bainum, Peter M.
2009-01-01
The Tschauner-Hempel equations are widely used to correct the separation distance drifts between a pair of satellites within a constellation in highly elliptical orbits [1]. This set of equations was discretized in the true anomaly angle [1] to be used in a digital steady-state hierarchical controller [2]. This controller [2] performed the drift correction between a pair of satellites within the constellation. The objective of a discretized system is to develop a simple algorithm to be implemented in the computer onboard the satellite. The main advantage of the discrete systems is that the computational time can be reduced by selecting a suitable sampling interval. For this digital system, the amount of data will depend on the sampling interval in the true anomaly angle [3]. The purpose of this paper is to implement the discrete Tschauner-Hempel equations and the steady-state hierarchical controller in the computer onboard the satellite. This set of equations is expressed in the true anomaly angle in which a relation will be formulated between the time and the true anomaly angle domains.
NASA Astrophysics Data System (ADS)
Wang, Ji; Fischer, Debra A.; Horch, Elliott P.; Xie, Ji-Wei
2015-06-01
As hundreds of gas giant planets have been discovered, we study how these planets form and evolve in different stellar environments, specifically in multiple stellar systems. In such systems, stellar companions may have a profound influence on gas giant planet formation and evolution via several dynamical effects such as truncation and perturbation. We select 84 Kepler Objects of Interest (KOIs) with gas giant planet candidates. We obtain high-angular resolution images using telescopes with adaptive optics (AO) systems. Together with the AO data, we use archival radial velocity data and dynamical analysis to constrain the presence of stellar companions. We detect 59 stellar companions around 40 KOIs for which we develop methods of testing their physical association. These methods are based on color information and galactic stellar population statistics. We find evidence of suppressive planet formation within 20 AU by comparing stellar multiplicity. The stellar multiplicity rate (MR) for planet host stars is {0}-0+5% within 20 AU. In comparison, the stellar MR is 18% ± 2% for the control sample, i.e., field stars in the solar neighborhood. The stellar MR for planet host stars is 34% ± 8% for separations between 20 and 200 AU, which is higher than the control sample at 12% ± 2%. Beyond 200 AU, stellar MRs are comparable between planet host stars and the control sample. We discuss the implications of the results on gas giant planet formation and evolution.
CRN5EXP: Expert system for statistical quality control
NASA Technical Reports Server (NTRS)
Hentea, Mariana
1991-01-01
The purpose of the Expert System CRN5EXP is to assist in checking the quality of the coils at two very important mills: Hot Rolling and Cold Rolling in a steel plant. The system interprets the statistical quality control charts, diagnoses and predicts the quality of the steel. Measurements of process control variables are recorded in a database and sample statistics such as the mean and the range are computed and plotted on a control chart. The chart is analyzed through patterns using the C Language Integrated Production System (CLIPS) and a forward chaining technique to reach a conclusion about the causes of defects and to take management measures for the improvement of the quality control techniques. The Expert System combines the certainty factors associated with the process control variables to predict the quality of the steel. The paper presents the approach to extract data from the database, the reason to combine certainty factors, the architecture and the use of the Expert System. However, the interpretation of control charts patterns requires the human expert's knowledge and lends to Expert Systems rules.
Visible light scatter measurements of the Advanced X-ray Astronomical Facility /AXAF/ mirror samples
NASA Technical Reports Server (NTRS)
Griner, D. B.
1981-01-01
NASA is studying the properties of mirror surfaces for X-ray telescopes, the data of which will be used to develop the telescope system for the Advanced X-ray Astronomical Facility. Visible light scatter measurements, using a computer controlled scanner, are made of various mirror samples to determine surface roughness. Total diffuse scatter is calculated using numerical integration techniques and used to estimate the rms surface roughness. The data measurements are then compared with X-ray scatter measurements of the same samples. A summary of the data generated is presented, along with graphs showing changes in scatter on samples before and after cleaning. Results show that very smooth surfaces can be polished on the common substrate materials (from 2 to 10 Angstroms), and nickel appears to give the lowest visible light scatter.
Irregular analytical errors in diagnostic testing - a novel concept.
Vogeser, Michael; Seger, Christoph
2018-02-23
In laboratory medicine, routine periodic analyses for internal and external quality control measurements interpreted by statistical methods are mandatory for batch clearance. Data analysis of these process-oriented measurements allows for insight into random analytical variation and systematic calibration bias over time. However, in such a setting, any individual sample is not under individual quality control. The quality control measurements act only at the batch level. Quantitative or qualitative data derived for many effects and interferences associated with an individual diagnostic sample can compromise any analyte. It is obvious that a process for a quality-control-sample-based approach of quality assurance is not sensitive to such errors. To address the potential causes and nature of such analytical interference in individual samples more systematically, we suggest the introduction of a new term called the irregular (individual) analytical error. Practically, this term can be applied in any analytical assay that is traceable to a reference measurement system. For an individual sample an irregular analytical error is defined as an inaccuracy (which is the deviation from a reference measurement procedure result) of a test result that is so high it cannot be explained by measurement uncertainty of the utilized routine assay operating within the accepted limitations of the associated process quality control measurements. The deviation can be defined as the linear combination of the process measurement uncertainty and the method bias for the reference measurement system. Such errors should be coined irregular analytical errors of the individual sample. The measurement result is compromised either by an irregular effect associated with the individual composition (matrix) of the sample or an individual single sample associated processing error in the analytical process. Currently, the availability of reference measurement procedures is still highly limited, but LC-isotope-dilution mass spectrometry methods are increasingly used for pre-market validation of routine diagnostic assays (these tests also involve substantial sets of clinical validation samples). Based on this definition/terminology, we list recognized causes of irregular analytical error as a risk catalog for clinical chemistry in this article. These issues include reproducible individual analytical errors (e.g. caused by anti-reagent antibodies) and non-reproducible, sporadic errors (e.g. errors due to incorrect pipetting volume due to air bubbles in a sample), which can both lead to inaccurate results and risks for patients.
On the Analysis of Case-Control Studies in Cluster-correlated Data Settings.
Haneuse, Sebastien; Rivera-Rodriguez, Claudia
2018-01-01
In resource-limited settings, long-term evaluation of national antiretroviral treatment (ART) programs often relies on aggregated data, the analysis of which may be subject to ecological bias. As researchers and policy makers consider evaluating individual-level outcomes such as treatment adherence or mortality, the well-known case-control design is appealing in that it provides efficiency gains over random sampling. In the context that motivates this article, valid estimation and inference requires acknowledging any clustering, although, to our knowledge, no statistical methods have been published for the analysis of case-control data for which the underlying population exhibits clustering. Furthermore, in the specific context of an ongoing collaboration in Malawi, rather than performing case-control sampling across all clinics, case-control sampling within clinics has been suggested as a more practical strategy. To our knowledge, although similar outcome-dependent sampling schemes have been described in the literature, a case-control design specific to correlated data settings is new. In this article, we describe this design, discuss balanced versus unbalanced sampling techniques, and provide a general approach to analyzing case-control studies in cluster-correlated settings based on inverse probability-weighted generalized estimating equations. Inference is based on a robust sandwich estimator with correlation parameters estimated to ensure appropriate accounting of the outcome-dependent sampling scheme. We conduct comprehensive simulations, based in part on real data on a sample of N = 78,155 program registrants in Malawi between 2005 and 2007, to evaluate small-sample operating characteristics and potential trade-offs associated with standard case-control sampling or when case-control sampling is performed within clusters.
NASA Technical Reports Server (NTRS)
Spivey, Reggie A.; Gilley, Scott; Ostrogorsky, Aleksander; Grugel, Richard; Smith, Guy; Luz, Paul
2003-01-01
The Solidification Using a Baffle in Sealed Ampoules (SUBSA) and Pore Formation and Mobility Investigation (PFMI) furnaces were developed for operation in the International Space Station (ISS) Microgravity Science Glovebox (MSG). Both furnaces were launched to the ISS on STS-111, June 4, 2002, and are currently in use on orbit. The SUBSA furnace provides a maximum temperature of 850 C and can accommodate a metal sample as large as 30 cm long and 12mm in diameter. SUBSA utilizes a gradient freeze process with a minimum cooldown rate of 0.5C per min, and a stability of +/- 0.15C. An 8 cm long transparent gradient zone coupled with a Cohu 3812 camera and quartz ampoule allows for observation and video recording of the solidification process. PFMI is a Bridgman type furnace that operates at a maximum temperature of 130C and can accommodate a sample 23cm long and 10mm in diameter. Two Cohu 3812 cameras mounted 90 deg apart move on a separate translation system which allows for viewing of the sample in the transparent hot zone and gradient zone independent of the furnace translation rate and direction. Translation rates for both the cameras and furnace can be specified from 0.5micrometers/sec to 100 micrometers/sec with a stability of +/-5%. The two furnaces share a Process Control Module (PCM) which controls the furnace hardware, a Data Acquisition Pad (DaqPad) which provides signal condition of thermal couple data, and two Cohu 3812 cameras. The hardware and software allow for real time monitoring and commanding of critical process control parameters. This paper will provide a detailed explanation of the SUBSA and PFMI systems along with performance data and some preliminary results from completed on-orbit processing runs.
Design of intelligent vehicle control system based on single chip microcomputer
NASA Astrophysics Data System (ADS)
Zhang, Congwei
2018-06-01
The smart car microprocessor uses the KL25ZV128VLK4 in the Freescale series of single-chip microcomputers. The image sampling sensor uses the CMOS digital camera OV7725. The obtained track data is processed by the corresponding algorithm to obtain track sideline information. At the same time, the pulse width modulation control (PWM) is used to control the motor and servo movements, and based on the digital incremental PID algorithm, the motor speed control and servo steering control are realized. In the project design, IAR Embedded Workbench IDE is used as the software development platform to program and debug the micro-control module, camera image processing module, hardware power distribution module, motor drive and servo control module, and then complete the design of the intelligent car control system.
NASA Astrophysics Data System (ADS)
Voronov, V. N.; Yegoshina, O. V.; Bolshakova, N. A.; Yarovoi, V. O.; Latt, Aie Min
2016-12-01
Typical disturbances in the dynamics of a corrective reagent dosing system under unsteady-state conditions during the unsatisfactory operation of a chemical control system with some water chemistry upsets at thermal and nuclear power stations are considered. An experimental setup representing a physical model for the water chemistry control system is described. The two disturbances, which are most frequently encountered in water chemistry control practice, such as a breakdown or shutdown of temperature compensation during pH measurement and an increase in the heat-transfer fluid flow rate, have been modeled in the process of study. The study of the effect produced by the response characteristics of chemical control analyzers on the operation of a reagent dosing system under unsteady-state conditions is important for the operative control of a water chemistry regime state. The effect of temperature compensation during pH measurement on the dynamics of an ammonia-dosing system in the manual and automatic cycle chemistry control modes has been studied. It has been demonstrated that the reading settling time of a pH meter in the manual ammonia- dosing mode grows with a breakdown in temperature compensation and a simultaneous increase in the temperature of a heat-transfer fluid sample. To improve the efficiency of water chemistry control, some systems for the quality control of a heat-transfer fluid by a chemical parameter with the obligatory compensation of a disturbance in its flow rate have been proposed for use. Experimental results will possibly differ from industrial data due to a great length of sampling lines. For this reason, corrective reagent dosing systems must be adapted to the conditions of a certain power-generating unit in the process of their implementation.
Optical instrument development for detection of pesticide residue in apple surface
NASA Astrophysics Data System (ADS)
Dhakal, Sagar; Li, Yongyu; Peng, Yankun; Chao, Kuanglin; Qin, Jianwei
2013-05-01
Apple is the world largest produced and consumed fruit item. At the same time, apple ranks number one among the fruit item contaminated with pesticide. This research focuses on development of laboratory based self-developed software and hardware for detection of commercially available organophosphorous pesticide (chlorpyrifos) in apple surface. A laser light source of 785nm was used to excite the sample, and Raman spectroscopy assembled with CCD camera was used for optical data acquisition. A hardware system was designed and fabricated to clamp and rotate apple sample of varying size maintaining constant working distance between optical probe and sample surface. Graphical Users Interface (GUI) based on LabView platform was developed to control the hardware system. The GUI was used to control the Raman system including CCD temperature, exposure time, track height and track centre, data acquisition, data processing and result prediction. Different concentrations of commercially available 48% chlorpyrifos pesticide solutions were prepared and gently placed in apple surface and dried. Raman spectral data at different points from same apple along the equatorial region were then acquired. The results show that prominent peaks at 341cm-1, 632cm-1 and 680 cm-1 represent the pesticide residue. The laboratory based experiment was able to detect pesticide solution of 20ppm within 3 seconds. A linear relation between Raman intensity and pesticide residue was developed with accuracy of 97.8%. The result of the research is promising and thus is a milestone for developing industrially desired real time, non-invasive pesticide residue detection technology in future.
A New Flying Wire System for the Tevatron
NASA Astrophysics Data System (ADS)
Blokland, Willem; Dey, Joseph; Vogel, Greg
1997-05-01
A new Flying Wires system replaces the old system to enhance the analysis of the beam emittance, improve the reliability, and handle the upcoming upgrades of the Tevatron. New VME data acquisition modules and timing modules allow for more bunches to be sampled more precisely. The programming language LabVIEW, running on a Macintosh computer, controls the VME modules and the nuLogic motion board that flies the wires. LabVIEW also analyzes and stores the data, and handles local and remote commands. The new system flies three wires and fits profiles of 72 bunches to a gaussian function within two seconds. A new console application operates the flying wires from any control console. This paper discusses the hardware and software setup, the capabilities and measurement results of the new Flying Wires system.
NASA Astrophysics Data System (ADS)
Kassamakov, Ivan; Maconi, Göran; Penttilä, Antti; Helander, Petteri; Gritsevich, Maria; Puranen, Tuomas; Salmi, Ari; Hæggström, Edward; Muinonen, Karri
2018-02-01
We present the design of a novel scatterometer for precise measurement of the angular Mueller matrix profile of a mm- to µm-sized sample held in place by sound. The scatterometer comprises a tunable multimode Argon-krypton laser (with possibility to set 1 of the 12 wavelengths in visible range), linear polarizers, a reference photomultiplier tube (PMT) for monitoring the beam intensity, and a micro-PMT module mounted radially towards the sample at an adjustable radius. The measurement angle is controlled by a motor-driven rotation stage with an accuracy of 15'. The system is fully automated using LabVIEW, including the FPGA-based data acquisition and the instrument's user interface. The calibration protocol ensures accurate measurements by using a control sphere sample (diameter 3 mm, refractive index of 1.5) fixed first on a static holder followed by accurate multi-wavelength measurements of the same sample levitated ultrasonically. To demonstrate performance of the scatterometer, we conducted detailed measurements of light scattered by a particle derived from the Chelyabinsk meteorite, as well as planetary analogue materials. The measurements are the first of this kind, since they are obtained using controlled spectral angular scattering including linear polarization effects, for arbitrary shaped objects. Thus, our novel approach permits a non-destructive, disturbance-free measurement with control of the orientation and location of the scattering object.
NASA Astrophysics Data System (ADS)
Arabshahi, P.; Chao, Y.; Chien, S.; Gray, A.; Howe, B. M.; Roy, S.
2008-12-01
In many areas of Earth science, including climate change research, there is a need for near real-time integration of data from heterogeneous and spatially distributed sensors, in particular in-situ and space- based sensors. The data integration, as provided by a smart sensor web, enables numerous improvements, namely, 1) adaptive sampling for more efficient use of expensive space-based sensing assets, 2) higher fidelity information gathering from data sources through integration of complementary data sets, and 3) improved sensor calibration. The specific purpose of the smart sensor web development presented here is to provide for adaptive sampling and calibration of space-based data via in-situ data. Our ocean-observing smart sensor web presented herein is composed of both mobile and fixed underwater in-situ ocean sensing assets and Earth Observing System (EOS) satellite sensors providing larger-scale sensing. An acoustic communications network forms a critical link in the web between the in-situ and space-based sensors and facilitates adaptive sampling and calibration. After an overview of primary design challenges, we report on the development of various elements of the smart sensor web. These include (a) a cable-connected mooring system with a profiler under real-time control with inductive battery charging; (b) a glider with integrated acoustic communications and broadband receiving capability; (c) satellite sensor elements; (d) an integrated acoustic navigation and communication network; and (e) a predictive model via the Regional Ocean Modeling System (ROMS). Results from field experiments, including an upcoming one in Monterey Bay (October 2008) using live data from NASA's EO-1 mission in a semi closed-loop system, together with ocean models from ROMS, are described. Plans for future adaptive sampling demonstrations using the smart sensor web are also presented.
IEDA Thesaurus: A Controlled Vocabulary for IEDA Systems to Advance Integration
NASA Astrophysics Data System (ADS)
Ji, P.; Lehnert, K. A.; Arko, R. A.; Song, L.; Hsu, L.; Carter, M. R.; Ferrini, V. L.; Ash, J.
2014-12-01
Integrated Earth Data Applications (IEDA) is a community-based facility that serves to support, sustain, and advance the geosciences by providing data services for observational geoscience data from the Ocean, Earth, and Polar Sciences. Many dedicated systems such as the Petrological Database (PetDB), Marine Geoscience Data System (MGDS), System for Earth Sample Registration (SESAR), Data Coordination Center for the U.S. Antarctic Program (USAP-DCC), etc., under the umbrella of the IEDA framework, were developed to support the preservation, discovery, retrieval, and analysis of a wide range of observational field and analytical data types from diverse communities. However, it is currently difficult to maintain consistency of indexing content within IEDA schema, and perform unified or precise searching of the data in these diverse systems as each system maintains separate vocabularies, hierarchies, authority files, or sub taxonomies. We present here the IEDA Thesaurus, a system, which combines existing separate controlled vocabularies from the different systems under the IEDA schema into a single master controlled vocabulary, also introducing some new top facets for future long-term use. The IEDA thesaurus contains structured terminology for petrology, geochemistry, sedimentology, oceanography, geochronology, and volcanology, and other general metadata fields. 18 top facets (also called 'top categories') are defined, including equipment, geographic gazetteer, geologic ages, geologic units, materials, etc. The terms of the thesaurus are cross validated with others popular geoscience vocabularies such as GeoRef Thesaurus, U.S. Geological Survey Library Classification System, Global Change Master Directory (GCMD), and Semantic Web for Earth and Environmental Terminology (SWEET) ontologies. The thesaurus is organized along with the ANSI/NISO Z39.19-2005 Guidelines for the Construction, Format, and Management of Monolingual Controlled Vocabularies, and is published using Simple Knowledge Organization System (SKOS) format. The IEDA thesaurus server provides classic web semantic features such as SPARQL, RESTful web services, and unique URI based on open source technologies.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wright, R.M.; Zander, M.E.; Brown, S.K.
1992-09-01
This paper describes the application of video image processing to beam profile measurements on the Ground Test Accelerator (GTA). A diagnostic was needed to measure beam profiles in the intermediate matching section (IMS) between the radio-frequency quadrupole (RFQ) and the drift tube linac (DTL). Beam profiles are measured by injecting puffs of gas into the beam. The light emitted from the beam-gas interaction is captured and processed by a video image processing system, generating the beam profile data. A general purpose, modular and flexible video image processing system, imagetool, was used for the GTA image profile measurement. The development ofmore » both software and hardware for imagetool and its integration with the GTA control system (GTACS) will be discussed. The software includes specialized algorithms for analyzing data and calibrating the system. The underlying design philosophy of imagetool was tested by the experience of building and using the system, pointing the way for future improvements. The current status of the system will be illustrated by samples of experimental data.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wright, R.M.; Zander, M.E.; Brown, S.K.
1992-01-01
This paper describes the application of video image processing to beam profile measurements on the Ground Test Accelerator (GTA). A diagnostic was needed to measure beam profiles in the intermediate matching section (IMS) between the radio-frequency quadrupole (RFQ) and the drift tube linac (DTL). Beam profiles are measured by injecting puffs of gas into the beam. The light emitted from the beam-gas interaction is captured and processed by a video image processing system, generating the beam profile data. A general purpose, modular and flexible video image processing system, imagetool, was used for the GTA image profile measurement. The development ofmore » both software and hardware for imagetool and its integration with the GTA control system (GTACS) will be discussed. The software includes specialized algorithms for analyzing data and calibrating the system. The underlying design philosophy of imagetool was tested by the experience of building and using the system, pointing the way for future improvements. The current status of the system will be illustrated by samples of experimental data.« less
NASA Astrophysics Data System (ADS)
Cook, Emily Jane
2008-12-01
This thesis presents the analysis of low angle X-ray scatter measurements taken with an energy dispersive system for substance identification, imaging and system control. Diffraction measurements were made on illicit drugs, which have pseudo- crystalline structures and thus produce diffraction patterns comprising a se ries of sharp peaks. Though the diffraction profiles of each drug are visually characteristic, automated detection systems require a substance identification algorithm, and multivariate analysis was selected as suitable. The software was trained with measured diffraction data from 60 samples covering 7 illicit drugs and 5 common cutting agents, collected with a range of statistical qual ities and used to predict the content of 7 unknown samples. In all cases the constituents were identified correctly and the contents predicted to within 15%. Soft tissues exhibit broad peaks in their diffraction patterns. Diffraction data were collected from formalin fixed breast tissue samples and used to gen erate images. Maximum contrast between healthy and suspicious regions was achieved using momentum transfer windows 1.04-1.10 and 1.84-1.90 nm_1. The resulting images had an average contrast of 24.6% and 38.9% compared to the corresponding transmission X-ray images (18.3%). The data was used to simulate the feedback for an adaptive imaging system and the ratio of the aforementioned momentum transfer regions found to be an excellent pa rameter. Investigation into the effects of formalin fixation on human breast tissue and animal tissue equivalents indicated that fixation in standard 10% buffered formalin does not alter the diffraction profiles of tissue in the mo mentum transfer regions examined, though 100% unbuffered formalin affects the profile of porcine muscle tissue (a substitute for glandular and tumourous tissue), though fat is unaffected.
Kelso, Kyle W.; Flocks, James G.
2015-01-01
Selection of the core site locations was based on geophysical surveys conducted around the islands from 2008 to 2010. The surveys, using acoustic systems to image and interpret the nearsurface stratigraphy, were conducted to investigate the geologic controls on island evolution. This data series serves as an archive of sediment data collected from August to September 2010, offshore of the Mississippi barrier islands. Data products, including descriptive core logs, core photographs, results of sediment grain-size analyses, sample location maps, and geographic information system (GIS) data files with accompanying formal Federal Geographic Data Committee (FDGC) metadata can be downloaded from the data products and downloads page.
Hybrid suboptimal control of multi-rate multi-loop sampled-data systems
NASA Technical Reports Server (NTRS)
Shieh, Leang S.; Chen, Gwangchywan; Tsai, Jason S. H.
1992-01-01
A hybrid state-space controller is developed for suboptimal digital control of multirate multiloop multivariable continuous-time systems. First, an LQR is designed for a continuous-time subsystem which has a large bandwidth and is connnected in the inner loop of the overall system. The designed LQR would optimally place the eigenvalues of a closed-loop subsystem in the common region of an open sector bounded by sector angles + or - pi/2k for k = 2 or 3 from the negative real axis and the left-hand side of a vertical line on the negative real axis in the s-plane. Then, the developed continuous-time state-feedback gain is converted into an equivalent fast-rate discrete-time state-feedback gain via a digital redesign technique (Tsai et al. 1989, Shieh et al. 1990) reviewed here. A real state reconstructor is redeveloped utilizing the fast-rate input-output data of the system of interest. The design procedure of multiloop multivariable systems using multirate samplers is shown, and a terminal homing missile system example is used to demonstrate the effectiveness of the proposed method.
Application of statistical process control to qualitative molecular diagnostic assays.
O'Brien, Cathal P; Finn, Stephen P
2014-01-01
Modern pathology laboratories and in particular high throughput laboratories such as clinical chemistry have developed a reliable system for statistical process control (SPC). Such a system is absent from the majority of molecular laboratories and where present is confined to quantitative assays. As the inability to apply SPC to an assay is an obvious disadvantage this study aimed to solve this problem by using a frequency estimate coupled with a confidence interval calculation to detect deviations from an expected mutation frequency. The results of this study demonstrate the strengths and weaknesses of this approach and highlight minimum sample number requirements. Notably, assays with low mutation frequencies and detection of small deviations from an expected value require greater sample numbers to mitigate a protracted time to detection. Modeled laboratory data was also used to highlight how this approach might be applied in a routine molecular laboratory. This article is the first to describe the application of SPC to qualitative laboratory data.
NASA Technical Reports Server (NTRS)
1994-01-01
The ChemScan UV-6100 is a spectrometry system originally developed by Biotronics Technologies, Inc. under a Small Business Innovation Research (SBIR) contract. It is marketed to the water and wastewater treatment industries, replacing "grab sampling" with on-line data collection. It analyzes the light absorbance characteristics of a water sample, simultaneously detects hundreds of individual wavelengths absorbed by chemical substances in a process solution, and quantifies the information. Spectral data is then processed by ChemScan analyzer and compared with calibration files in the system's memory in order to calculate concentrations of chemical substances that cause UV light absorbance in specific patterns. Monitored substances can be analyzed for quality and quantity. Applications include detection of a variety of substances, and the information provided enables an operator to control a process more efficiently.
Statistical transformation and the interpretation of inpatient glucose control data.
Saulnier, George E; Castro, Janna C; Cook, Curtiss B
2014-03-01
To introduce a statistical method of assessing hospital-based non-intensive care unit (non-ICU) inpatient glucose control. Point-of-care blood glucose (POC-BG) data from hospital non-ICUs were extracted for January 1 through December 31, 2011. Glucose data distribution was examined before and after Box-Cox transformations and compared to normality. Different subsets of data were used to establish upper and lower control limits, and exponentially weighted moving average (EWMA) control charts were constructed from June, July, and October data as examples to determine if out-of-control events were identified differently in nontransformed versus transformed data. A total of 36,381 POC-BG values were analyzed. In all 3 monthly test samples, glucose distributions in nontransformed data were skewed but approached a normal distribution once transformed. Interpretation of out-of-control events from EWMA control chart analyses also revealed differences. In the June test data, an out-of-control process was identified at sample 53 with nontransformed data, whereas the transformed data remained in control for the duration of the observed period. Analysis of July data demonstrated an out-of-control process sooner in the transformed (sample 55) than nontransformed (sample 111) data, whereas for October, transformed data remained in control longer than nontransformed data. Statistical transformations increase the normal behavior of inpatient non-ICU glycemic data sets. The decision to transform glucose data could influence the interpretation and conclusions about the status of inpatient glycemic control. Further study is required to determine whether transformed versus nontransformed data influence clinical decisions or evaluation of interventions.
Furukawa, Makoto; Takagai, Yoshitaka
2016-10-04
Online solid-phase extraction (SPE) coupled with inductively coupled plasma mass spectrometry (ICPMS) is a useful tool in automatic sequential analysis. However, it cannot simultaneously quantify the analytical targets and their recovery percentages (R%) in one-shot samples. We propose a system that simultaneously acquires both data in a single sample injection. The main flowline of the online solid-phase extraction is divided into main and split flows. The split flow line (i.e., bypass line), which circumvents the SPE column, was placed on the main flow line. Under program-controlled switching of the automatic valve, the ICPMS sequentially measures the targets in a sample before and after column preconcentration and determines the target concentrations and the R% on the SPE column. This paper describes the system development and two demonstrations to exhibit the analytical significance, i.e., the ultratrace amounts of radioactive strontium ( 90 Sr) using commercial Sr-trap resin and multielement adsorbability on the SPE column. This system is applicable to other flow analyses and detectors in online solid phase extraction.
Dual-view plane illumination microscopy for rapid and spatially isotropic imaging
Kumar, Abhishek; Wu, Yicong; Christensen, Ryan; Chandris, Panagiotis; Gandler, William; McCreedy, Evan; Bokinsky, Alexandra; Colón-Ramos, Daniel A; Bao, Zhirong; McAuliffe, Matthew; Rondeau, Gary; Shroff, Hari
2015-01-01
We describe the construction and use of a compact dual-view inverted selective plane illumination microscope (diSPIM) for time-lapse volumetric (4D) imaging of living samples at subcellular resolution. Our protocol enables a biologist with some prior microscopy experience to assemble a diSPIM from commercially available parts, to align optics and test system performance, to prepare samples, and to control hardware and data processing with our software. Unlike existing light sheet microscopy protocols, our method does not require the sample to be embedded in agarose; instead, samples are prepared conventionally on glass coverslips. Tissue culture cells and Caenorhabditis elegans embryos are used as examples in this protocol; successful implementation of the protocol results in isotropic resolution and acquisition speeds up to several volumes per s on these samples. Assembling and verifying diSPIM performance takes ~6 d, sample preparation and data acquisition take up to 5 d and postprocessing takes 3–8 h, depending on the size of the data. PMID:25299154
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
An Apple IIe microcomputer is being used to collect data and to control a pyrolysis system. Pyrolysis data for bitumen and kerogen are widely used to estimate source rock maturity. For a detailed analysis of kinetic parameters, however, data must be obtained more precisely than for routine pyrolysis. The authors discuss the program which controls the temperature ramp of the furnace that heats the sample, and collects data from a thermocouple in the furnace and from the flame ionization detector measuring evolved hydrocarbons. These data are stored on disk for later use by programs that display the results of themore » experiment or calculate kinetic parameters. The program is written in Applesoft BASIC with subroutines in Apple assembler for speed and efficiency.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Holladay, S.K.; Anderson, H.M.; Benson, S.B.
Quality assurance (QA) objectives for Phase 2 were that (1) scientific data generated would withstand scientific and legal scrutiny; (2) data would be gathered using appropriate procedures for sample collection, sample handling and security, chain of custody, laboratory analyses, and data reporting; (3) data would be of known precision and accuracy; and (4) data would meet data quality objectives defined in the Phase 2 Sampling and Analysis Plan. A review of the QA systems and quality control (QC) data associated with the Phase 2 investigation is presented to evaluate whether the data were of sufficient quality to satisfy Phase 2more » objectives. The data quality indicators of precision, accuracy, representativeness, comparability, completeness, and sensitivity were evaluated to determine any limitations associated with the data. Data were flagged with qualifiers that were associated with appropriate reason codes and documentation relating the qualifiers to the reviewer of the data. These qualifiers were then consolidated into an overall final qualifier to represent the quality of the data to the end user. In summary, reproducible, precise, and accurate measurements consistent with CRRI objectives and the limitations of the sampling and analytical procedures used were obtained for the data collected in support of the Phase 2 Remedial Investigation.« less
Practical aspects of modeling aircraft dynamics from flight data
NASA Technical Reports Server (NTRS)
Iliff, K. W.; Maine, R. E.
1984-01-01
The purpose of parameter estimation, a subset of system identification, is to estimate the coefficients (such as stability and control derivatives) of the aircraft differential equations of motion from sampled measured dynamic responses. In the past, the primary reason for estimating stability and control derivatives from flight tests was to make comparisons with wind tunnel estimates. As aircraft became more complex, and as flight envelopes were expanded to include flight regimes that were not well understood, new requirements for the derivative estimates evolved. For many years, the flight determined derivatives were used in simulations to aid in flight planning and in pilot training. The simulations were particularly important in research flight test programs in which an envelope expansion into new flight regimes was required. Parameter estimation techniques for estimating stability and control derivatives from flight data became more sophisticated to support the flight test programs. As knowledge of these new flight regimes increased, more complex aircraft were flown. Much of this increased complexity was in sophisticated flight control systems. The design and refinement of the control system required higher fidelity simulations than were previously required.
Light curves of flat-spectrum radio sources (Jenness+, 2010)
NASA Astrophysics Data System (ADS)
Jenness, T.; Robson, E. I.; Stevens, J. A.
2010-05-01
Calibrated data for 143 flat-spectrum extragalactic radio sources are presented at a wavelength of 850um covering a 5-yr period from 2000 April. The data, obtained at the James Clerk Maxwell Telescope using the Submillimetre Common-User Bolometer Array (SCUBA) camera in pointing mode, were analysed using an automated pipeline process based on the Observatory Reduction and Acquisition Control - Data Reduction (ORAC-DR) system. This paper describes the techniques used to analyse and calibrate the data, and presents the data base of results along with a representative sample of the better-sampled light curves. A re-analysis of previously published data from 1997 to 2000 is also presented. The combined catalogue, comprising 10493 flux density measurements, provides a unique and valuable resource for studies of extragalactic radio sources. (2 data files).
Serious injury prediction algorithm based on large-scale data and under-triage control.
Nishimoto, Tetsuya; Mukaigawa, Kosuke; Tominaga, Shigeru; Lubbe, Nils; Kiuchi, Toru; Motomura, Tomokazu; Matsumoto, Hisashi
2017-01-01
The present study was undertaken to construct an algorithm for an advanced automatic collision notification system based on national traffic accident data compiled by Japanese police. While US research into the development of a serious-injury prediction algorithm is based on a logistic regression algorithm using the National Automotive Sampling System/Crashworthiness Data System, the present injury prediction algorithm was based on comprehensive police data covering all accidents that occurred across Japan. The particular focus of this research is to improve the rescue of injured vehicle occupants in traffic accidents, and the present algorithm assumes the use of an onboard event data recorder data from which risk factors such as pseudo delta-V, vehicle impact location, seatbelt wearing or non-wearing, involvement in a single impact or multiple impact crash and the occupant's age can be derived. As a result, a simple and handy algorithm suited for onboard vehicle installation was constructed from a sample of half of the available police data. The other half of the police data was applied to the validation testing of this new algorithm using receiver operating characteristic analysis. An additional validation was conducted using in-depth investigation of accident injuries in collaboration with prospective host emergency care institutes. The validated algorithm, named the TOYOTA-Nihon University algorithm, proved to be as useful as the US URGENCY and other existing algorithms. Furthermore, an under-triage control analysis found that the present algorithm could achieve an under-triage rate of less than 10% by setting a threshold of 8.3%. Copyright © 2016 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Hou, Ligang; Luo, Rengui; Wu, Wuchen
2006-11-01
This paper forwards a low power grating detection chip (EYAS) on length and angle precision measurement. Traditional grating detection method, such as resister chain divide or phase locked divide circuit are difficult to design and tune. The need of an additional CPU for control and display makes these methods' implementation more complex and costly. Traditional methods also suffer low sampling speed for the complex divide circuit scheme and CPU software compensation. EYAS is an application specific integrated circuit (ASIC). It integrates micro controller unit (MCU), power management unit (PMU), LCD controller, Keyboard interface, grating detection unit and other peripherals. Working at 10MHz, EYAS can afford 5MHz internal sampling rate and can handle 1.25MHz orthogonal signal from grating sensor. With a simple control interface by keyboard, sensor parameter, data processing and system working mode can be configured. Two LCD controllers can adapt to dot array LCD or segment bit LCD, which comprised output interface. PMU alters system between working and standby mode by clock gating technique to save power. EYAS in test mode (system action are more frequently than real world use) consumes 0.9mw, while 0.2mw in real world use. EYAS achieved the whole grating detection system function, high-speed orthogonal signal handling in a single chip with very low power consumption.
Espinosa, Manuel; Weinberg, Diego; Rotela, Camilo H; Polop, Francisco; Abril, Marcelo; Scavuzzo, Carlos Marcelo
2016-05-01
Since 2009, Fundación Mundo Sano has implemented an Aedes aegypti Surveillance and Control Program in Tartagal city (Salta Province, Argentina). The purpose of this study was to analyze temporal dynamics of Ae. aegypti breeding sites spatial distribution, during five years of samplings, and the effect of control actions over vector population dynamics. Seasonal entomological (larval) samplings were conducted in 17,815 fixed sites in Tartagal urban area between 2009 and 2014. Based on information of breeding sites abundance, from satellite remote sensing data (RS), and by the use of Geographic Information Systems (GIS), spatial analysis (hotspots and cluster analysis) and predictive model (MaxEnt) were performed. Spatial analysis showed a distribution pattern with the highest breeding densities registered in city outskirts. The model indicated that 75% of Ae. aegypti distribution is explained by 3 variables: bare soil coverage percentage (44.9%), urbanization coverage percentage(13.5%) and water distribution (11.6%). This results have called attention to the way entomological field data and information from geospatial origin (RS/GIS) are used to infer scenarios which could then be applied in epidemiological surveillance programs and in the determination of dengue control strategies. Predictive maps development constructed with Ae. aegypti systematic spatiotemporal data, in Tartagal city, would allow public health workers to identify and target high-risk areas with appropriate and timely control measures. These tools could help decision-makers to improve health system responses and preventive measures related to vector control.
Espinosa, Manuel; Weinberg, Diego; Rotela, Camilo H.; Polop, Francisco; Abril, Marcelo; Scavuzzo, Carlos Marcelo
2016-01-01
Background Since 2009, Fundación Mundo Sano has implemented an Aedes aegypti Surveillance and Control Program in Tartagal city (Salta Province, Argentina). The purpose of this study was to analyze temporal dynamics of Ae. aegypti breeding sites spatial distribution, during five years of samplings, and the effect of control actions over vector population dynamics. Methodology/Principal Findings Seasonal entomological (larval) samplings were conducted in 17,815 fixed sites in Tartagal urban area between 2009 and 2014. Based on information of breeding sites abundance, from satellite remote sensing data (RS), and by the use of Geographic Information Systems (GIS), spatial analysis (hotspots and cluster analysis) and predictive model (MaxEnt) were performed. Spatial analysis showed a distribution pattern with the highest breeding densities registered in city outskirts. The model indicated that 75% of Ae. aegypti distribution is explained by 3 variables: bare soil coverage percentage (44.9%), urbanization coverage percentage(13.5%) and water distribution (11.6%). Conclusions/Significance This results have called attention to the way entomological field data and information from geospatial origin (RS/GIS) are used to infer scenarios which could then be applied in epidemiological surveillance programs and in the determination of dengue control strategies. Predictive maps development constructed with Ae. aegypti systematic spatiotemporal data, in Tartagal city, would allow public health workers to identify and target high-risk areas with appropriate and timely control measures. These tools could help decision-makers to improve health system responses and preventive measures related to vector control. PMID:27223693
SNIFFER: An aerial platform for the plume phase of a nuclear emergency
NASA Astrophysics Data System (ADS)
Castelluccio, D. M.; Cisbani, E.; Frullani, S.
2012-04-01
When a nuclear or radiological accident results in a release of a radioactive plume, AGS (Aerial Gamma Spectrometry) systems used in many countries, equipped with passive detectors, can help in giving quantitative assessment on the radiological situation (land surface contamination level) only when the air contamination due to the passage of the travelling plume has become negligible. To overcome this limitation, the Italian Institute of Health has developed and implemented a multi purpose air sampling system based on a fixed wing aircraft, for time-effective, large areas radiological surveillance (to face radiological emergency and to support homeland security). A fixed wing aircraft (Sky Arrow 650) with the front part of the fuselage properly adapted to house the detection equipment has been equipped with a compact air sampling line where the isokinetic sampling is dynamically maintained. Aerosol is collected on a Teflon® filter positioned along the line and hosted on a rotating 4-filters disk. A complex of detectors allows radionuclide identification in the collected aerosol samples. A correlated analysis of these two detectors data allows a quantitative measurement of air as well as ground surface concentration of gamma emitting radioisotopes. Environmental sensors and a GPS receiver support the characterization of the sampling conditions and the temporal and geolocation of the acquired data. Acquisition and control system based on compact electronics and real time software that operate the sampling line actuators, guarantee the dynamical isokinetic condition, and acquire the detectors and sensor data. The system is also equipped with other sampling lines to provide information on the concentration of other chemical pollutants. Operative flights have been carried out in the last years, and performances and results are presented.
Development and Flight Testing of an Autonomous Landing Gear Health-Monitoring System
NASA Technical Reports Server (NTRS)
Woodard, Stanley E.; Coffey, Neil C.; Gonzalez, Guillermo A.; Taylor, B. Douglas; Brett, Rube R.; Woodman, Keith L.; Weathered, Brenton W.; Rollins, Courtney H.
2003-01-01
Development and testing of an adaptable vehicle health-monitoring architecture is presented. The architecture is being developed for a fleet of vehicles. It has three operational levels: one or more remote data acquisition units located throughout the vehicle; a command and control unit located within the vehicle; and, a terminal collection unit to collect analysis results from all vehicles. Each level is capable of performing autonomous analysis with a trained expert system. Communication between all levels is done with wireless radio frequency interfaces. The remote data acquisition unit has an eight channel programmable digital interface that allows the user discretion for choosing type of sensors; number of sensors, sensor sampling rate and sampling duration for each sensor. The architecture provides framework for a tributary analysis. All measurements at the lowest operational level are reduced to provide analysis results necessary to gauge changes from established baselines. These are then collected at the next level to identify any global trends or common features from the prior level. This process is repeated until the results are reduced at the highest operational level. In the framework, only analysis results are forwarded to the next level to reduce telemetry congestion. The system's remote data acquisition hardware and non-analysis software have been flight tested on the NASA Langley B757's main landing gear. The flight tests were performed to validate the following: the wireless radio frequency communication capabilities of the system, the hardware design, command and control; software operation; and, data acquisition, storage and retrieval.
Noise in NC-AFM measurements with significant tip–sample interaction
Lübbe, Jannis; Temmen, Matthias
2016-01-01
The frequency shift noise in non-contact atomic force microscopy (NC-AFM) imaging and spectroscopy consists of thermal noise and detection system noise with an additional contribution from amplitude noise if there are significant tip–sample interactions. The total noise power spectral density D Δ f(f m) is, however, not just the sum of these noise contributions. Instead its magnitude and spectral characteristics are determined by the strongly non-linear tip–sample interaction, by the coupling between the amplitude and tip–sample distance control loops of the NC-AFM system as well as by the characteristics of the phase locked loop (PLL) detector used for frequency demodulation. Here, we measure D Δ f(f m) for various NC-AFM parameter settings representing realistic measurement conditions and compare experimental data to simulations based on a model of the NC-AFM system that includes the tip–sample interaction. The good agreement between predicted and measured noise spectra confirms that the model covers the relevant noise contributions and interactions. Results yield a general understanding of noise generation and propagation in the NC-AFM and provide a quantitative prediction of noise for given experimental parameters. We derive strategies for noise-optimised imaging and spectroscopy and outline a full optimisation procedure for the instrumentation and control loops. PMID:28144538
Noise in NC-AFM measurements with significant tip-sample interaction.
Lübbe, Jannis; Temmen, Matthias; Rahe, Philipp; Reichling, Michael
2016-01-01
The frequency shift noise in non-contact atomic force microscopy (NC-AFM) imaging and spectroscopy consists of thermal noise and detection system noise with an additional contribution from amplitude noise if there are significant tip-sample interactions. The total noise power spectral density D Δ f ( f m ) is, however, not just the sum of these noise contributions. Instead its magnitude and spectral characteristics are determined by the strongly non-linear tip-sample interaction, by the coupling between the amplitude and tip-sample distance control loops of the NC-AFM system as well as by the characteristics of the phase locked loop (PLL) detector used for frequency demodulation. Here, we measure D Δ f ( f m ) for various NC-AFM parameter settings representing realistic measurement conditions and compare experimental data to simulations based on a model of the NC-AFM system that includes the tip-sample interaction. The good agreement between predicted and measured noise spectra confirms that the model covers the relevant noise contributions and interactions. Results yield a general understanding of noise generation and propagation in the NC-AFM and provide a quantitative prediction of noise for given experimental parameters. We derive strategies for noise-optimised imaging and spectroscopy and outline a full optimisation procedure for the instrumentation and control loops.
NASA Technical Reports Server (NTRS)
Lebel, P. J.; Lamontagne, R. A.; Goldstein, H. W.
1976-01-01
The Carbon Monoxide Pollution Experiment (COPE) and the National Capital Air Quality Control Region (NCAQCR) undertook a series of measurements of atmospheric CO and CH4 to determine the accuracy of the airborne COPE Correlation Interfer4meter. The device, a modified Michelson interferometer, measures the atmospheric column density of CO and CH4 at 2.3 microns with tropospheric measurement sensitivities of 70 and 10 PPB, respectively. Data for evaluating the remote measurements included atmospheric column density measurements at a ground truth site using a van-mounted infrared Fourier spectrometer; continuous ground level gas chromatographic measurements; and chromatographic data from atmospheric grab samples collected by aircraft and at ground locations. The instruments and sampling techniques used in the experiment are described in detail.
A wireless high-speed data acquisition system for geotechnical centrifuge model testing
NASA Astrophysics Data System (ADS)
Gaudin, C.; White, D. J.; Boylan, N.; Breen, J.; Brown, T.; DeCatania, S.; Hortin, P.
2009-09-01
This paper describes a novel high-speed wireless data acquisition system (WDAS) developed at the University of Western Australia for operation onboard a geotechnical centrifuge, in an enhanced gravitational field of up to 300 times Earth's gravity. The WDAS system consists of up to eight separate miniature units distributed around the circumference of a 0.8 m diameter drum centrifuge, communicating with the control room via wireless Ethernet. Each unit is capable of powering and monitoring eight instrument channels at a sampling rate of up to 1 MHz at 16-bit resolution. The data are stored within the logging unit in solid-state memory, but may also be streamed in real-time at low frequency (up to 10 Hz) to the centrifuge control room, via wireless transmission. The high-speed logging runs continuously within a circular memory (buffer), allowing for storage of a pre-trigger segment of data prior to an event. To suit typical geotechnical modelling applications, the system can record low-speed data continuously, until a burst of high-speed acquisition is triggered when an experimental event occurs, after which the system reverts back to low-speed acquisition to monitor the aftermath of the event. Unlike PC-based data acquisition solutions, this system performs the full sequence of amplification, conditioning, digitization and storage on a single circuit board via an independent micro-controller allocated to each pair of instrumented channels. This arrangement is efficient, compact and physically robust to suit the centrifuge environment. This paper details the design specification of the WDAS along with the software interface developed to control the units. Results from a centrifuge test of a submarine landslide are used to illustrate the performance of the new WDAS.
On modeling of integrated communication and control systems
NASA Technical Reports Server (NTRS)
Liou, Luen-Woei; Ray, Asok
1990-01-01
The mathematical modeling scheme proposed by Ray and Halevi (1988) for integrated communication and control systems is considered analytically, with an emphasis on the effect of introducing varying and distributed time delays to account for asynchronous time-division multiplexing in the communication part of the system. Ray and Halevi applied a state-transition concept to transform the original continuous-time model into a discrete-time model; the same approach was used by Kalman and Bertram (1959) to model various types of sampled data systems which are not subject to induced delays. The relationship between the two modeling schemes is explored, and it is shown that, although the Kalman-Bertram method has the advantage of a unified approach, it becomes inconvenient when varying delays appear in the control loop.
Guo, Ye; Chen, Qian; Wu, Wei; Cui, Wei
2015-03-31
To establish a system of monitoring the key indicator of quality for inspection (KIQI) on a laboratory information system (LIS), and to have a better management of KIQI. Clinical sample made in PUMCH were collected during the whole of 2014. Next, interactive input program were designed to accomplish data collecting of the disqualification rate of samples, the mistake rate of samples and the occasions of losing samples, etc. Then, a series moment of sample collection, laboratory sample arrived, sample test, sample check, response to critical value, namely, trajectory information left on LIS were recorded and the qualification rate of TAT, the notification rate of endangering result were calculated. Finally, the information about quality control were collected to build an internal quality control database and the KIQI, such as the out-of-control rate of quality control and the total error of test items were monitored. The inspection of the sample management shows the disqualification rates in 2014 were all below the target, but the rates in January and February were a little high and the rates of four wards were above 2%. The mistake rates of samples was 0.47 cases/10 000 cases, attaining the target (< 2 cases/10 000 cases). Also, there was no occasion of losing samples in 2014, attaining the target too. The inspection of laboratory reports shows the qualification rates of TAT was within the acceptable range (> 95%), however the rates of blood routine in November (94.75%) was out of range. We have solved the problem by optimizing the processes. The notification rate of endangering result attained the target (≥ 98%), while the rate of timely notification is needed to improve. Quality inspection shows the CV of APTT in August (5.02%) was rising significantly, beyond the accepted CV (5.0%). We have solved the problem by changing the reagent. The CV of TT in 2014 were all below the allowable CV, thus the allowable CV of the next year lower to 10%. It is an objective and effective method to manage KIQI with the powerful management mode of database and information process capability on LIS.
7 CFR 90.2 - General terms defined.
Code of Federal Regulations, 2011 CFR
2011-01-01
... agency, or other agency, organization or person that defines in the general terms the basis on which the... analytical data using proficiency check sample or analyte recovery techniques. In addition, the certainty.... Quality control. The system of close examination of the critical details of an analytical procedure in...
DOE Office of Scientific and Technical Information (OSTI.GOV)
DeRosier, R.; Waterland, L.R.
1987-03-01
The report gives emission results from field tests of a wood-waste-fired industrial firetube boiler. Emission measurements included: continuous monitoring of flue-gas emissions; source assessment sampling system (SASS) sampling of the flue gas with subsequent laboratory analysis of samples to give total flue-gas organics in two boiling-point ranges, compound category information within these ranges, specific quantitation of the semivolatile organic priority pollutants, and flue-gas concentrations of 65 trace elements; Method 5 sampling for particulates; controlled condensation system (CSS) sampling for SO/sub 2/ and SO/sub 3/; and grab sampling of boiler bottom ash for trace-element-content determinations. Emission levels of five polycyclic organicmore » matter species and phenol were quantitated: except for naphthalene, all were emitted at less than 0.4 microgram/dscm.« less
Noise and complexity in human postural control: interpreting the different estimations of entropy.
Rhea, Christopher K; Silver, Tobin A; Hong, S Lee; Ryu, Joong Hyun; Studenka, Breanna E; Hughes, Charmayne M L; Haddad, Jeffrey M
2011-03-17
Over the last two decades, various measures of entropy have been used to examine the complexity of human postural control. In general, entropy measures provide information regarding the health, stability and adaptability of the postural system that is not captured when using more traditional analytical techniques. The purpose of this study was to examine how noise, sampling frequency and time series length influence various measures of entropy when applied to human center of pressure (CoP) data, as well as in synthetic signals with known properties. Such a comparison is necessary to interpret data between and within studies that use different entropy measures, equipment, sampling frequencies or data collection durations. The complexity of synthetic signals with known properties and standing CoP data was calculated using Approximate Entropy (ApEn), Sample Entropy (SampEn) and Recurrence Quantification Analysis Entropy (RQAEn). All signals were examined at varying sampling frequencies and with varying amounts of added noise. Additionally, an increment time series of the original CoP data was examined to remove long-range correlations. Of the three measures examined, ApEn was the least robust to sampling frequency and noise manipulations. Additionally, increased noise led to an increase in SampEn, but a decrease in RQAEn. Thus, noise can yield inconsistent results between the various entropy measures. Finally, the differences between the entropy measures were minimized in the increment CoP data, suggesting that long-range correlations should be removed from CoP data prior to calculating entropy. The various algorithms typically used to quantify the complexity (entropy) of CoP may yield very different results, particularly when sampling frequency and noise are different. The results of this study are discussed within the context of the neural noise and loss of complexity hypotheses.
Neuromorphic learning of continuous-valued mappings from noise-corrupted data
NASA Technical Reports Server (NTRS)
Troudet, T.; Merrill, W.
1991-01-01
The effect of noise on the learning performance of the backpropagation algorithm is analyzed. A selective sampling of the training set is proposed to maximize the learning of control laws by backpropagation, when the data have been corrupted by noise. The training scheme is applied to the nonlinear control of a cart-pole system in the presence of noise. The neural computation provides the neurocontroller with good noise-filtering properties. In the presence of plant noise, the neurocontroller is found to be more stable than the teacher. A novel perspective on the application of neural network technology to control engineering is presented.
Decrease of oxidative phosphorylation system function in severe septic patients.
Lorente, Leonardo; Martín, María M; López-Gallardo, Ester; Blanquer, José; Solé-Violán, Jordi; Labarta, Lorenzo; Díaz, César; Jiménez, Alejandro; Montoya, Julio; Ruiz-Pesini, Eduardo
2015-10-01
The comparison of oxidative phosphorylation system capacities between septic patients and control subjects has been scarcely analyzed and only in studies with small sample size (fewer than 40 septic patients and 40 controls). Thus, the objective of this study was to compare platelet respiratory complex IV (CIV) activity between severe septic patients and healthy individuals in a larger series (including 198 severe septic patients and 96 healthy controls). A prospective, multicenter, observational study was carried out in 6 Spanish intensive care units. We obtained blood samples from 198 severe septic patients at day 1, 4, and 8 of the severe sepsis diagnosis and 96 sex- and age-matched healthy control individuals and determined platelet CIV-specific activity. The end point of the study was 30-day mortality. Control individuals showed higher platelet CIV-specific activity (P < .001) than surviving (n = 130) or nonsurviving (n = 68) severe septic patients at day 1, 4, and 8 of severe sepsis diagnosis. The major finding of our work, involving the largest series to date of severe septic patients with data on oxidative phosphorylation system capacity, was that surviving and nonsurviving septic patients showed lower platelet CIV-specific activity during the first week of sepsis than healthy controls. Copyright © 2015 Elsevier Inc. All rights reserved.
Fine PM measurements: personal and indoor air monitoring.
Jantunen, M; Hänninen, O; Koistinen, K; Hashim, J H
2002-12-01
This review compiles personal and indoor microenvironment particulate matter (PM) monitoring needs from recently set research objectives, most importantly the NRC published "Research Priorities for Airborne Particulate Matter (1998)". Techniques and equipment used to monitor PM personal exposures and microenvironment concentrations and the constituents of the sampled PM during the last 20 years are then reviewed. Development objectives are set and discussed for personal and microenvironment PM samplers and monitors, for filter materials, and analytical laboratory techniques for equipment calibration, filter weighing and laboratory climate control. The progress is leading towards smaller sample flows, lighter, silent, independent (battery powered) monitors with data logging capacity to store microenvironment or activity relevant sensor data, advanced flow controls and continuous recording of the concentration. The best filters are non-hygroscopic, chemically pure and inert, and physically robust against mechanical wear. Semiautomatic and primary standard equivalent positive displacement flow meters are replacing the less accurate methods in flow calibration, and also personal sampling flow rates should become mass flow controlled (with or without volumetric compensation for pressure and temperature changes). In the weighing laboratory the alternatives are climatic control (set temperature and relative humidity), and mechanically simpler thermostatic heating, air conditioning and dehumidification systems combined with numerical control of temperature, humidity and pressure effects on flow calibration and filter weighing.
Maljovec, D.; Liu, S.; Wang, B.; ...
2015-07-14
Here, dynamic probabilistic risk assessment (DPRA) methodologies couple system simulator codes (e.g., RELAP and MELCOR) with simulation controller codes (e.g., RAVEN and ADAPT). Whereas system simulator codes model system dynamics deterministically, simulation controller codes introduce both deterministic (e.g., system control logic and operating procedures) and stochastic (e.g., component failures and parameter uncertainties) elements into the simulation. Typically, a DPRA is performed by sampling values of a set of parameters and simulating the system behavior for that specific set of parameter values. For complex systems, a major challenge in using DPRA methodologies is to analyze the large number of scenarios generated,more » where clustering techniques are typically employed to better organize and interpret the data. In this paper, we focus on the analysis of two nuclear simulation datasets that are part of the risk-informed safety margin characterization (RISMC) boiling water reactor (BWR) station blackout (SBO) case study. We provide the domain experts a software tool that encodes traditional and topological clustering techniques within an interactive analysis and visualization environment, for understanding the structures of such high-dimensional nuclear simulation datasets. We demonstrate through our case study that both types of clustering techniques complement each other for enhanced structural understanding of the data.« less
An Analysis of U.S. Army Health Hazard Assessments During the Acquisition of Military Materiel
2010-06-03
protective equipment (PPE) (Milz, Conrad, & Soule , 2003). Engineering controls can eliminate hazards through system design, substitution of hazardous...Milz, Conrad, & Soule , 2003). Engineering control measures can serve to 7 minimize hazards where they cannot be eliminated, with preference for...during the materiel acquisitions process, and (c) will evaluate a sample of the database for accuracy by comparing the data entries to original reports
Mansano, Raul K; Godoy, Eduardo P; Porto, Arthur J V
2014-12-18
Recent advances in wireless networking technology and the proliferation of industrial wireless sensors have led to an increasing interest in using wireless networks for closed loop control. The main advantages of Wireless Networked Control Systems (WNCSs) are the reconfigurability, easy commissioning and the possibility of installation in places where cabling is impossible. Despite these advantages, there are two main problems which must be considered for practical implementations of WNCSs. One problem is the sampling period constraint of industrial wireless sensors. This problem is related to the energy cost of the wireless transmission, since the power supply is limited, which precludes the use of these sensors in several closed-loop controls. The other technological concern in WNCS is the energy efficiency of the devices. As the sensors are powered by batteries, the lowest possible consumption is required to extend battery lifetime. As a result, there is a compromise between the sensor sampling period, the sensor battery lifetime and the required control performance for the WNCS. This paper develops a model-based soft sensor to overcome these problems and enable practical implementations of WNCSs. The goal of the soft sensor is generating virtual data allowing an actuation on the process faster than the maximum sampling period available for the wireless sensor. Experimental results have shown the soft sensor is a solution to the sampling period constraint problem of wireless sensors in control applications, enabling the application of industrial wireless sensors in WNCSs. Additionally, our results demonstrated the soft sensor potential for implementing energy efficient WNCS through the battery saving of industrial wireless sensors.
Lefkowitz, Daniel; Pechter, Elise; Fitzsimmons, Kathleen; Lumia, Margaret; Stephens, Alicia C.; Davis, Letitia; Flattery, Jennifer; Weinberg, Justine; Harrison, Robert J.; Reilly, Mary Jo; Filios, Margaret S.; White, Gretchen E.; Rosenman, Kenneth D.
2015-01-01
Background Isocyanates remain a leading cause of work-related asthma (WRA). Methods Two independent data systems were analyzed for the period 1993–2008: (1) State-based WRA case surveillance data on persons with isocyanate-induced WRA from four states, and (2) Occupational Safety and Health Administration (OSHA) Integrated Management Information System (IMIS) isocyanate air sampling results. Results We identified 368 cases of isocyanate-induced WRA from 32 industries and 678 OSHA isocyanate air samples with detectable levels from 31 industries. Seventeen industries were unique to one or the other dataset. Conclusion Isocyanate-induced WRA continues to occur in a wide variety of industries. Two data systems uncovered industries with isocyanate exposures and/or illness. Improved control measures and standards, including medical surveillance, are needed. More emphasis is needed on task-specific guidance, spill clean-up procedures, skin and respiratory protection, and targeted medical monitoring to mitigate the hazards of isocyanate use. PMID:26351141
Development of an Oculometer Data Collection Subsystem.
1982-12-01
number of time increments between 50 and 183 . "Data gone" is a data sample with an instrument number of zero and the total number of time increments...greater than 183 . Table 2 shows the organization of the information in the data tables. Table 3 shows the format of each of the rows. A ’$’ in front of a...APPENDIX A Memory FFFF SYM-1 Operating System Locations EFFF FF80 Resident Assembler/Editor DFFF CFFF BFFF Resident Assembler/Editor AFFF Disk controller
NASA Technical Reports Server (NTRS)
James, John T.
2001-01-01
The toxicological assessment of air samples returned at the end of the STS-102 (5A.1) flight to the ISS is reported. ISS air samples were taken in late February 2001 from the Service Module, FGB, and U.S. Laboratory using grab sample canisters (GSCs) and/or formaldehyde badges . A "first-entry" sample of the multipurpose logistics module (MPLM) atmosphere was taken with a GSC, and preflight and end-of-mission samples were obtained from Discovery using GSCs. Analytical methods have not changed from earlier reports, and all quality control measures were met for the data presented herein. The two general criteria used to assess air quality are the total-non-methane-volatile organic hydrocarbons (NMVOCs) and the total T-value (minus the CO2 contribution). Control of atmospheric alcohols is important to the water recovery system engineers, hence total alcohols were also assessed in each sample. Formaldehyde is quantified separately.
Vortex boundary-layer interactions
NASA Technical Reports Server (NTRS)
Bradshaw, P.
1986-01-01
Parametric studies to identify a vortex generator were completed. Data acquisition in the first chosen configuration, in which a longitudinal vortex pair generated by an isolated delta wing starts to merge with a turbulent boundary layer on a flat plate fairly close to the leading edge is nearly completed. Work on a delta-wing/flat-plate combination, consisting of a flow visualization and hot wire measurements taken with a computer controlled traverse gear and data logging system were completed. Data taking and analysis have continued, and sample results for another cross stream plane are presented. Available data include all mean velocity components, second order mean products of turbulent fluctuations, and third order mean products. Implementation of a faster data logging system was accomplished.
Cancer control needs of 2-1-1 callers in Missouri, North Carolina, Texas, and Washington.
Purnell, Jason Q; Kreuter, Matthew W; Eddens, Katherine S; Ribisl, Kurt M; Hannon, Peggy; Williams, Rebecca S; Fernandez, Maria E; Jobe, David; Gemmel, Susan; Morris, Marti; Fagin, Debbie
2012-05-01
Innovative interventions are needed to connect underserved populations to cancer control services. With data from Missouri, North Carolina, Texas, and Washington this study a) estimated the cancer control needs of callers to 2-1-1, an information and referral system used by underserved populations, b) compared rates of need with state and national data, and c) examined receptiveness to needed referrals. From October 2009 to March 2010 callers' (N=1,408) cancer control needs were assessed in six areas: breast, cervical, and colorectal cancer screening, HPV vaccination, smoking, and smoke-free homes using Behavioral Risk Factor Surveillance System (BRFSS) survey items. Standardized estimates were compared with state and national rates. Nearly 70% of the sample had at least one cancer control need. Needs were greater for 2-1-1 callers than for state and national rates, and callers were receptive to referrals. 2-1-1 could be a key partner in efforts to reduce cancer disparities.
Oblinger, Carolyn J.
2004-01-01
The Triangle Area Water Supply Monitoring Project was initiated in October 1988 to provide long-term water-quality data for six area water-supply reservoirs and their tributaries. In addition, the project provides data that can be used to determine the effectiveness of large-scale changes in water-resource management practices, document differences in water quality among water-supply types (large multiuse reservoir, small reservoir, run-of-river), and tributary-loading and in-lake data for water-quality modeling of Falls and Jordan Lakes. By September 2001, the project had progressed in four phases and included as many as 34 sites (in 1991). Most sites were sampled and analyzed by the U.S. Geological Survey. Some sites were already a part of the North Carolina Division of Water Quality statewide ambient water-quality monitoring network and were sampled by the Division of Water Quality. The network has provided data on streamflow, physical properties, and concentrations of nutrients, major ions, metals, trace elements, chlorophyll, total organic carbon, suspended sediment, and selected synthetic organic compounds. Project quality-assurance activities include written procedures for sample collection, record management and archive, collection of field quality-control samples (blank samples and replicate samples), and monitoring the quality of field supplies. In addition to project quality-assurance activities, the quality of laboratory analyses was assessed through laboratory quality-assurance practices and an independent laboratory quality-control assessment provided by the U.S. Geological Survey Branch of Quality Systems through the Blind Inorganic Sample Project and the Organic Blind Sample Project.
Quality-control design for surface-water sampling in the National Water-Quality Network
Riskin, Melissa L.; Reutter, David C.; Martin, Jeffrey D.; Mueller, David K.
2018-04-10
The data-quality objectives for samples collected at surface-water sites in the National Water-Quality Network include estimating the extent to which contamination, matrix effects, and measurement variability affect interpretation of environmental conditions. Quality-control samples provide insight into how well the samples collected at surface-water sites represent the true environmental conditions. Quality-control samples used in this program include field blanks, replicates, and field matrix spikes. This report describes the design for collection of these quality-control samples and the data management needed to properly identify these samples in the U.S. Geological Survey’s national database.
25th Space Simulation Conference. Environmental Testing: The Earth-Space Connection
NASA Technical Reports Server (NTRS)
Packard, Edward
2008-01-01
Topics covered include: Methods of Helium Injection and Removal for Heat Transfer Augmentation; The ESA Large Space Simulator Mechanical Ground Support Equipment for Spacecraft Testing; Temperature Stability and Control Requirements for Thermal Vacuum/Thermal Balance Testing of the Aquarius Radiometer; The Liquid Nitrogen System for Chamber A: A Change from Original Forced Flow Design to a Natural Flow (Thermo Siphon) System; Return to Mercury: A Comparison of Solar Simulation and Flight Data for the MESSENGER Spacecraft; Floating Pressure Conversion and Equipment Upgrades of Two 3.5kw, 20k, Helium Refrigerators; Affect of Air Leakage into a Thermal-Vacuum Chamber on Helium Refrigeration Heat Load; Special ISO Class 6 Cleanroom for the Lunar Reconnaissance Orbiter (LRO) Project; A State-of-the-Art Contamination Effects Research and Test Facility Martian Dust Simulator; Cleanroom Design Practices and Their Influence on Particle Counts; Extra Terrestrial Environmental Chamber Design; Contamination Sources Effects Analysis (CSEA) - A Tool to Balance Cost/Schedule While Managing Facility Availability; SES and Acoustics at GSFC; HST Super Lightweight Interchangeable Carrier (SLIC) Static Test; Virtual Shaker Testing: Simulation Technology Improves Vibration Test Performance; Estimating Shock Spectra: Extensions beyond GEVS; Structural Dynamic Analysis of a Spacecraft Multi-DOF Shaker Table; Direct Field Acoustic Testing; Manufacture of Cryoshroud Surfaces for Space Simulation Chambers; The New LOTIS Test Facility; Thermal Vacuum Control Systems Options for Test Facilities; Extremely High Vacuum Chamber for Low Outgassing Processing at NASA Goddard; Precision Cleaning - Path to Premier; The New Anechoic Shielded Chambers Designed for Space and Commercial Applications at LIT; Extraction of Thermal Performance Values from Samples in the Lunar Dust Adhesion Bell Jar; Thermal (Silicon Diode) Data Acquisition System; Aquarius's Instrument Science Data System (ISDS) Automated to Acquire, Process, Trend Data and Produce Radiometric System Assessment Reports; Exhaustive Thresholds and Resistance Checkpoints; Reconfigurable HIL Testing of Earth Satellites; FPGA Control System for the Automated Test of MicroShutters; Ongoing Capabilities and Developments of Re-Entry Plasma Ground Tests at EADS-ASTRIUM; Operationally Responsive Space Standard Bus Battery Thermal Balance Testing and Heat Dissipation Analysis; Galileo - The Serial-Production AIT Challenge; The Space Systems Environmental Test Facility Database (SSETFD), Website Development Status; Simulated Reentry Heating by Torching; Micro-Vibration Measurements on Thermally Loaded Multi-Layer Insulation Samples in Vacuum; High Temperature Life Testing of 80Ni-20Cr Wire in a Simulated Mars Atmosphere for the Sample Analysis at Mars (SAM) Instrument Suit Gas Processing System (GPS) Carbon Dioxide Scrubber; The Planning and Implementation of Test Facility Improvements; and Development of a Silicon Carbide Molecular Beam Nozzle for Simulation Planetary Flybys and Low-Earth Orbit.
NASA Astrophysics Data System (ADS)
Tong, Kai; Fan, Shiming; Gong, Derong; Lu, Zuming; Liu, Jian
The synchronizer/data buffer (SDB) in the command and data acquisition station for China's future Geostationary Meteorological Satellite is described. Several computers and special microprocessors are used in tandem with minimized hardware to fulfill all of the functions. The high-accuracy digital phase locked loop is operated by computer and by controlling the count value of the 20-MHz clock to acquire and track such signals as sun pulse, scan synchronization detection pulse, and earth pulse. Sun pulse and VISSR data are recorded precisely and economically by digitizing the time relation. The VISSR scan timing and equiangular control timing, and equal time sampling on satellite are also discussed.
NASA Astrophysics Data System (ADS)
Nakatani, T.; Inamura, Y.; Moriyama, K.; Ito, T.; Muto, S.; Otomo, T.
Neutron scattering can be a powerful probe in the investigation of many phenomena in the materials and life sciences. The Materials and Life Science Experimental Facility (MLF) at the Japan Proton Accelerator Research Complex (J-PARC) is a leading center of experimental neutron science and boasts one of the most intense pulsed neutron sources in the world. The MLF currently has 18 experimental instruments in operation that support a wide variety of users from across a range of research fields. The instruments include optical elements, sample environment apparatus and detector systems that are controlled and monitored electronically throughout an experiment. Signals from these components and those from the neutron source are converted into a digital format by the data acquisition (DAQ) electronics and recorded as time-tagged event data in the DAQ computers using "DAQ-Middleware". Operating in event mode, the DAQ system produces extremely large data files (˜GB) under various measurement conditions. Simultaneously, the measurement meta-data indicating each measurement condition is recorded in XML format by the MLF control software framework "IROHA". These measurement event data and meta-data are collected in the MLF common storage and cataloged by the MLF Experimental Database (MLF EXP-DB) based on a commercial XML database. The system provides a web interface for users to manage and remotely analyze experimental data.
Initial operation of the NSTX-Upgrade real-time velocity diagnostic
Podestà, M.; Bell, R. E.
2016-11-03
A real-time velocity (RTV) diagnostic based on active charge-exchange recombination spectroscopy is now operational on the National Spherical Torus Experiment-Upgrade (NSTX-U) spherical torus (Menard et al 2012 Nucl. Fusion 52 083015). We designed the system in order to supply plasma velocity data in real time to the NSTX-U plasma control system, as required for the implementation of toroidal rotation control. Our measurements are available from four radii at a maximum sampling frequency of 5 kHz. Post-discharge analysis of RTV data provides additional information on ion temperature, toroidal velocity and density of carbon impurities. Furthermore, examples of physics studies enabled bymore » RTV measurements from initial operations of NSTX-U are discussed.« less
Intelligent process control of fiber chemical vapor deposition
NASA Astrophysics Data System (ADS)
Jones, John Gregory
Chemical Vapor Deposition (CVD) is a widely used process for the application of thin films. In this case, CVD is being used to apply a thin film interface coating to single crystal monofilament sapphire (Alsb2Osb3) fibers for use in Ceramic Matrix Composites (CMC's). The hot-wall reactor operates at near atmospheric pressure which is maintained using a venturi pump system. Inert gas seals obviate the need for a sealed system. A liquid precursor delivery system has been implemented to provide precise stoichiometry control. Neural networks have been implemented to create real-time process description models trained using data generated based on a Navier-Stokes finite difference model of the process. Automation of the process to include full computer control and data logging capability is also presented. In situ sensors including a quadrupole mass spectrometer, thermocouples, laser scanner, and Raman spectrometer have been implemented to determine the gas phase reactants and coating quality. A fuzzy logic controller has been developed to regulate either the gas phase or the in situ temperature of the reactor using oxygen flow rate as an actuator. Scanning electron microscope (SEM) images of various samples are shown. A hierarchical control structure upon which the control structure is based is also presented.
Valid statistical inference methods for a case-control study with missing data.
Tian, Guo-Liang; Zhang, Chi; Jiang, Xuejun
2018-04-01
The main objective of this paper is to derive the valid sampling distribution of the observed counts in a case-control study with missing data under the assumption of missing at random by employing the conditional sampling method and the mechanism augmentation method. The proposed sampling distribution, called the case-control sampling distribution, can be used to calculate the standard errors of the maximum likelihood estimates of parameters via the Fisher information matrix and to generate independent samples for constructing small-sample bootstrap confidence intervals. Theoretical comparisons of the new case-control sampling distribution with two existing sampling distributions exhibit a large difference. Simulations are conducted to investigate the influence of the three different sampling distributions on statistical inferences. One finding is that the conclusion by the Wald test for testing independency under the two existing sampling distributions could be completely different (even contradictory) from the Wald test for testing the equality of the success probabilities in control/case groups under the proposed distribution. A real cervical cancer data set is used to illustrate the proposed statistical methods.
Bacteriological etiology and treatment of mastitis in Finnish dairy herds.
Vakkamäki, Johanna; Taponen, Suvi; Heikkilä, Anna-Maija; Pyörälä, Satu
2017-05-25
The Finnish dairy herd recording system maintains production and health records of cows and herds. Veterinarians and farmers register veterinary treatments in the system. Milk samples for microbiological analysis are routinely taken from mastitic cows. The laboratory of the largest dairy company in Finland, Valio Ltd., analyzes most samples using real-time PCR. This study addressed pathogen-specific microbiological data and treatment and culling records, in combination with cow and herd characteristics, from the Finnish dairy herd recording system during 2010-2012. The data derived from 240,067 quarter milk samples from 93,529 dairy cows with mastitis; 238,235 cows from the same herds served as the control group. No target pathogen DNA was detected in 12% of the samples. In 49% of the positive samples, only one target species and in 19%, two species with one dominant species were present. The most common species in the samples with a single species only were coagulase-negative staphylococci (CNS) (43%), followed by Staphylococcus aureus (21%), Streptococcus uberis (9%), Streptococcus dysgalactiae (8%), Corynebacterium bovis (7%), and Escherichia coli (5%). On average, 36% of the study cows and 6% of the control cows had recorded mastitis treatments during lactation. The corresponding proportions were 16 and 6% at drying-off. For more than 75% of the treatments during lactation, diagnosis was acute clinical mastitis. In the milk samples from cows with a recorded mastitis treatment during lactation, CNS and S. aureus were most common, followed by streptococci. Altogether, 48% of the cows were culled during the study. Mastitis was reported as the most common reason to cull; 49% of study cows and 18% of control cows were culled because of mastitis. Culling was most likely if S. aureus was detected in the milk sample submitted during the culling year. The PCR test has proven to be an applicable method also for large-scale use in bacterial diagnostics. In the present study, microbiological diagnosis was unequivocal in the great majority of samples where a single species or two species with one dominating were detected. Coagulase-negative staphylococci and S. aureus were the most common species. S. aureus was also the most common pathogen among the culled cows, which emphasizes the importance of preventive measures.
Global Tobacco Surveillance System (GTSS): purpose, production, and potential.
2005-01-01
The World Health Organization (WHO), Centers for Disease Control and Prevention (CDC), and Canadian Public Health Association (CPHA) developed the Global Tobacco Surveillance System (GTSS) to assist all 192 WHO Member States in collecting data on youth and adult tobacco use. The flexible GTSS system includes common data items but allows countries to include important unique information at their discretion. It uses a common survey methodology, similar field procedures for data collection, and similar data management and processing techniques. The GTSS includes collection of data through three surveys: the Global Youth Tobacco Survey (GYTS) for youth, and the Global School Personnel Survey (GSPS) and the Global Health Professional Survey (GHPS) for adults. GTSS data potentially can be applied in four ways. First, countries and research partners can disseminate data through publications, presentations, and an active GTSS web site. Second, countries can use GTSS data to inform politicians about the tobacco problem in their country, leading to new policy decisions to prevent and control tobacco use. Third, GTSS can provide countries with valuable feedback to evaluate and improve Country National Action Plans or develop new plans. Fourth, in response to the WHO FCTC call for countries to use consistent methods and procedures in their surveillance efforts, GTSS offers such consistency in sampling procedures, core questionnaire items, training infield procedures, and analysis of data across all survey sites. The GTSS represents the most comprehensive tobacco surveillance system ever developed and implemented. As an example, this paper describes development of the GYTS and discusses potential uses of the data. Sample data were drawn from 38 sites in 24 countries in the African Region, 82 sites in 35 countries in the Americas Region, 20 sites in 17 countries and the Gaza Strip/West Bank region in the Eastern Mediterranean Region, 25 sites in 22 countries in the European Region, 34 sites in six countries in the Southeast Asia Region, and 25 sites in 14 countries in the Western Pacific Region.
Stereo Sound Field Controller Design Using Partial Model Matching on the Frequency Domain
NASA Astrophysics Data System (ADS)
Kumon, Makoto; Miike, Katsuhiro; Eguchi, Kazuki; Mizumoto, Ikuro; Iwai, Zenta
The objective of sound field control is to make the acoustic characteristics of a listening room close to those of the desired system. Conventional methods apply feedforward controllers, such as digital filters, to achieve this objective. However, feedback controllers are also necessary in order to attenuate noise or to compensate the uncertainty of the acoustic characteristics of the listening room. Since acoustic characteristics are well modeled on the frequency domain, it is efficient to design controllers with respect to frequency responses, but it is difficult to design a multi input multi output (MIMO) control system on a wide frequency domain. In the present study, a partial model matching method on the frequency domain was adopted because this method requires only sampled data, rather than complex mathematical models of the plant, in order to design controllers for MIMO systems. The partial model matching method was applied to design two-degree-of-freedom controllers for acoustic equalization and noise reduction. Experiments demonstrated effectiveness of the proposed method.
NASA Astrophysics Data System (ADS)
Sani, S.; Saad, M. H. Md; Jamaludin, N.; Ismail, M. P.; Mohd, S.; Mustapha, I.; Masenwat, N. A.; Tengku Amran, T. S.; Megat Ahmad, M. H. A.
2018-01-01
This paper discussed the design and development of a portable PC-based ultrasonic goniometer system that can be used to study material properties using ultrasonic wave. The system utilizes an ultrasonic pulse-receiver card model attached to computer notebook for signal display. A new specific software package (GoNIO) was developed to control the operation of the scanner, displaying the data and analyze characteristics of materials. System testing was carried out using samples with cubic dimension of about 10 mm x 20 mm x 30 mm. This size allows the sample to be fitted into the goniometer specimen holder and immersed in a liquid during measurement. The sample was rotated from incident angle of 0° to 90° during measurement and the amplitude reflected signals were recorded at every one degree of rotation. Immersion transducers were used to generate and receive the ultrasounds that pass through the samples. Longitudinal, shear and Rayleigh wave measurements were performed on the samples to determine the Dynamic Young’s Modulus. Results of measurements are explained and discussed.
Biological monitoring of glazers exposed to lead in the ceramics industry in Iran.
Shouroki, Fatemeh Kargar; Shahtaheri, Seyed Jamaleddin; Golbabaei, Farideh; Barkhordari, Abolfazle; Rahimi-Froushani, Abbas
2015-01-01
Exposure to heavy metals, particularly lead, takes place in the ceramics industry. Lead is used in glaze to produce smooth and brilliant surfaces; thus, there is a likelihood of occupational adverse effects on humans. Urine samples were collected from 49 glazers at the start and end of the work shifts (98 samples). Solid phase extraction was used for separation and pre-concentration of the analyte. Samples were analysed by inductively coupled plasma-atomic emission spectroscopy (ICP-AES). Lung function tests were performed on both control and lead exposed subjects. Statistical analysis of covariance (ANCOVA) was used to evaluate the data obtained. The concentration of lead in glazers was 6.37 times higher than in the control group. Lung functions were significantly lower in the glazers compared to the control group (p < 0.001). Results showed that poor ventilation systems, overtime work and work history are effective determinants of high exposure levels.
In-line quality control of moving objects by means of spectral-domain OCT
NASA Astrophysics Data System (ADS)
Markl, Daniel; Hannesschläger, Günther; Buchsbaum, Andreas; Sacher, Stephan; Khinast, Johannes G.; Leitner, Michael
2014-08-01
In-line quality control of intermediate and final products is essential in various industries. This may imply determining the thickness of a foil or evaluating the homogeneity of coating applied to a pharmaceutical tablet. Such a qualitative and quantitative monitoring in a depth-resolved manner can be accomplished using optical coherence tomography (OCT). In-line quality control based on OCT requires additional consideration of motion effects for the system design as well as for data interpretation. This study focuses on transverse motion effects that can arise in spectral-domain (SD-) OCT systems. The impact of a transverse movement is analyzed for a constant relative speed difference up to 0.7 m/s between sample and sensor head. In particular, transverse motion is affecting OCT system properties such as the beam displacement (distance between adjacent A-scans) and transverse resolution. These properties were evaluated theoretically and experimentally for OCT images of a resolution target and pharmaceutical film-coated tablets. Both theoretical and experimental analyses highlight the shift of the transverse resolution limiting factor from the optics to the beam displacement above a relative speed difference between sensor head and sample of 0.42 m/s (for the presented SD-OCT setup). Speeds above 0.4 m/s are often demanded when monitoring industrial processes, such as a coating process when producing film-coated tablets. This emphasizes the importance of a fast data acquisition when using OCT as in-line quality control tool.
Lloyd, T L; Perschy, T B; Gooding, A E; Tomlinson, J J
1992-01-01
A fully automated assay for the analysis of ranitidine in serum and plasma, with and without an internal standard, was validated. It utilizes robotic solid phase extraction with on-line high performance liquid chromatographic (HPLC) analysis. The ruggedness of the assay was demonstrated over a three-year period. A Zymark Py Technology II robotic system was used for serial processing from initial aspiration of samples from original collection containers, to final direct injection onto the on-line HPLC system. Automated serial processing with on-line analysis provided uniform sample history and increased productivity by freeing the chemist to analyse data and perform other tasks. The solid phase extraction efficiency was 94% throughout the assay range of 10-250 ng/mL. The coefficients of variation for within- and between-day quality control samples ranged from 1 to 6% and 1 to 5%, respectively. Mean accuracy for between-day standards and quality control results ranged from 97 to 102% of the respective theoretical concentrations.
Waschbusch, Robert J.
2003-01-01
The Wisconsin Department of Transportation is required to control the quality of runoff from roadways under their control as part of the National Pollution Discharge Elimination System. One way to control roadway runoff is to use street sweeping to remove pollutants before they are entrained in runoff. This may be a good option because land is often unavailable or prohibitively expensive and structural best-management practices can also be expensive. This study collected stormwater runoff samples and dirt samples from the roadway surface from a section of Interstate Highway 894 near Milwaukee, Wisconsin during periods when a street sweeping program was and was not in effect. These data may be useful in evaluating street sweeping as a stormwater best management practice but this study did not perform this evaluation. Data collection methods, concentrations of sediment and other constituents in storm- water runoff, and street dirt masses are presented in this report. Replicate and comparison sample results indicate that when evaluating the effectiveness of best-management practices on highway runoff, suspended sediment results should be used rather than suspended solids, presumably because the particle sizes in highway runoff is large compared to those found in other types of stormwater runoff.
WRATS Integrated Data Acquisition System
NASA Technical Reports Server (NTRS)
Piatak, David J.
2008-01-01
The Wing and Rotor Aeroelastic Test System (WRATS) data acquisition system (DAS) is a 64-channel data acquisition display and analysis system specifically designed for use with the WRATS 1/5-scale V-22 tiltrotor model of the Bell Osprey. It is the primary data acquisition system for experimental aeroelastic testing of the WRATS model for the purpose of characterizing the aeromechanical and aeroelastic stability of prototype tiltrotor configurations. The WRATS DAS was also used during aeroelastic testing of Bell Helicopter Textron s Quad-Tiltrotor (QTR) design concept, a test which received international attention. The LabVIEW-based design is portable and capable of powering and conditioning over 64 channels of dynamic data at sampling rates up to 1,000 Hz. The system includes a 60-second circular data archive, an integrated model swashplate excitation system, a moving block damping application for calculation of whirl flutter mode subcritical damping, a loads and safety monitor, a pilot-control console display, data analysis capabilities, and instrumentation calibration functions. Three networked computers running custom-designed LabVIEW software acquire data through National Instruments data acquisition hardware. The aeroelastic model (see figure) was tested with the DAS at two facilities at NASA Langley, the Transonic Dynamics Tunnel (TDT) and the Rotorcraft Hover Test Facility (RHTF). Because of the need for seamless transition between testing at these facilities, DAS is portable. The software is capable of harmonic analysis of periodic time history data, Fast Fourier Transform calculations, power spectral density calculations, and on-line calibration of test instrumentation. DAS has a circular buffer archive to ensure critical data is not lost in event of model failure/incident, as well as a sample-and-hold capability for phase-correct time history data.
A three-dimensional orthogonal laser velocimeter for the NASA Ames 7- by 10-foot wind tunnel
NASA Technical Reports Server (NTRS)
Dunagan, Stephen E.; Cooper, Donald L.
1995-01-01
A three-component dual-beam laser-velocimeter system has been designed, fabricated, and implemented in the 7-by 10-Foot Wind Tunnel at NASA Ames Research Center. The instrument utilizes optical access from both sides and the top of the test section, and is configured for uncoupled orthogonal measurements of the three Cartesian coordinates of velocity. Bragg cell optics are used to provide fringe velocity bias. Modular system design provides great flexibility in the location of sending and receiving optics to adapt to specific experimental requirements. Near-focus Schmidt-Cassegrain optic modules may be positioned for collection of forward or backward scattered light over a large solid angle, and may be clustered to further increase collection solid angle. Multimode fiber optics transmit collected light to the photomultiplier tubes for processing. Counters are used to process the photomultiplier signals and transfer the processed data digitally via buffered interface controller to the host MS-DOS computer. Considerable data reduction and graphical display programming permit on-line control of data acquisition and evaluation of the incoming data. This paper describes this system in detail and presents sample data illustrating the system's capability.
Converting the Active Digital Controller for Use in Two Tests
NASA Technical Reports Server (NTRS)
Wright, Robert G.
1995-01-01
The Active Digital Controller is a system used to control the various functions of wind tunnel models. It has the capability of digitizing and saving of up to sixty-four channels of analog data. It can output up to 16 channels of analog command signals. In addition to its use as a general controller, it can run up to two distinct control laws. All of this is done at a regulated speed of two hundred hertz. The Active Digital Controller (ADC) was modified for use in the Actively Controlled Response of Buffet Affected Tails (ACROBAT) tests and for side-wall pressure data acquisition. The changes included general maintenance and updating of the controller as well as setting up special modes of operation. The ACROBAT tests required that two sets of output signals be available. The pressure data acquisition needed a sampling rate of four hundred hertz, twice the standard ADC rate. These modifications were carried out and the ADC was used during the ACROBAT wind tunnel entry.
Versatile, low-cost, computer-controlled, sample positioning system for vacuum applications
NASA Technical Reports Server (NTRS)
Vargas-Aburto, Carlos; Liff, Dale R.
1991-01-01
A versatile, low-cost, easy to implement, microprocessor-based motorized positioning system (MPS) suitable for accurate sample manipulation in a Second Ion Mass Spectrometry (SIMS) system, and for other ultra-high vacuum (UHV) applications was designed and built at NASA LeRC. The system can be operated manually or under computer control. In the latter case, local, as well as remote operation is possible via the IEEE-488 bus. The position of the sample can be controlled in three linear orthogonal and one angular coordinates.
Spatial characterization of the meltwater field from icebergs in the Weddell Sea.
Helly, John J; Kaufmann, Ronald S; Vernet, Maria; Stephenson, Gordon R
2011-04-05
We describe the results from a spatial cyberinfrastructure developed to characterize the meltwater field around individual icebergs and integrate the results with regional- and global-scale data. During the course of the cyberinfrastructure development, it became clear that we were also building an integrated sampling planning capability across multidisciplinary teams that provided greater agility in allocating expedition resources resulting in new scientific insights. The cyberinfrastructure-enabled method is a complement to the conventional methods of hydrographic sampling in which the ship provides a static platform on a station-by-station basis. We adapted a sea-floor mapping method to more rapidly characterize the sea surface geophysically and biologically. By jointly analyzing the multisource, continuously sampled biological, chemical, and physical parameters, using Global Positioning System time as the data fusion key, this surface-mapping method enables us to examine the relationship between the meltwater field of the iceberg to the larger-scale marine ecosystem of the Southern Ocean. Through geospatial data fusion, we are able to combine very fine-scale maps of dynamic processes with more synoptic but lower-resolution data from satellite systems. Our results illustrate the importance of spatial cyberinfrastructure in the overall scientific enterprise and identify key interfaces and sources of error that require improved controls for the development of future Earth observing systems as we move into an era of peta- and exascale, data-intensive computing.
Spatial characterization of the meltwater field from icebergs in the Weddell Sea
Helly, John J.; Kaufmann, Ronald S.; Vernet, Maria; Stephenson, Gordon R.
2011-01-01
We describe the results from a spatial cyberinfrastructure developed to characterize the meltwater field around individual icebergs and integrate the results with regional- and global-scale data. During the course of the cyberinfrastructure development, it became clear that we were also building an integrated sampling planning capability across multidisciplinary teams that provided greater agility in allocating expedition resources resulting in new scientific insights. The cyberinfrastructure-enabled method is a complement to the conventional methods of hydrographic sampling in which the ship provides a static platform on a station-by-station basis. We adapted a sea-floor mapping method to more rapidly characterize the sea surface geophysically and biologically. By jointly analyzing the multisource, continuously sampled biological, chemical, and physical parameters, using Global Positioning System time as the data fusion key, this surface-mapping method enables us to examine the relationship between the meltwater field of the iceberg to the larger-scale marine ecosystem of the Southern Ocean. Through geospatial data fusion, we are able to combine very fine-scale maps of dynamic processes with more synoptic but lower-resolution data from satellite systems. Our results illustrate the importance of spatial cyberinfrastructure in the overall scientific enterprise and identify key interfaces and sources of error that require improved controls for the development of future Earth observing systems as we move into an era of peta- and exascale, data-intensive computing. PMID:21444769
[Cardiovascular circulation feedback control treatment instrument].
Ge, Yu-zhi; Zhu, Xing-huan; Sheng, Guo-tai; Cao, Ping-liang; Liu, Dong-sheng; Wu, Zhi-ting
2005-07-01
The cardiovascular circulation feedback control treatment instrument (CFCTI) is an automatic feedback control treatment system, which has the function of monitoring, alarming, trouble self-diagnosis and testing on the line in the closed loop. The instrument is designed based on the successful clinical experiences and the data are inputted into the computer in real-time through a pressure sensor and A/D card. User interface window is set up for the doctor's choosing different medicine. The orders are outputted to control the dose of medicine through the transfusion system. The response to medicine is updated continually. CFCTI can avoid the man-made errors and the long interval of sampling. Its reliability and accuracy in rescuing the critical patients are much higher than the traditional methods.
NASA Technical Reports Server (NTRS)
Parra, Macarena; Jung, Jimmy; Almeida, Eduardo; Boone, Travis; Schonfeld, Julie; Tran, Luan
2016-01-01
The WetLab-2 system was developed by NASA Ames Research Center to offer new capabilities to researchers. The system can lyse cells and extract RNA (Ribonucleic Acid) on-orbit from different sample types ranging from microbial cultures to animal tissues. The purified RNA can then either be stabilized for return to Earth or can be used to conduct on-orbit quantitative Reverse Transcriptase PCR (Polymerase Chain Reaction) (qRT-PCR) analysis without the need for sample return. The qRT-PCR results can be downlinked to the ground a few hours after the completion of the run. The validation flight of the WetLab-2 system launched on SpaceX-8 on April 8, 2016. On orbit operations started on April 15th with system setup and was followed by three quantitative PCR runs using an E. coli genomic DNA template pre-loaded at three different concentrations. These runs were designed to discern if quantitative PCR functions correctly in microgravity and if the data is comparable to that from the ground control runs. The flight data showed no significant differences compared to the ground data though there was more variability in the values, this was likely due to the numerous small bubbles observed. The capability of the system to process samples and purify RNA was then validated using frozen samples prepared on the ground. The flight data for both E. coli and mouse liver clearly shows that RNA was successfully purified by our system. The E. coli qRT-PCR run showed successful singleplex, duplex and triplex capability. Data showed high variability in the resulting Cts (Cycle Thresholds [for the PCR]) likely due to bubble formation and insufficient mixing during the procedure run. The mouse liver qRT-PCR run had successful singleplex and duplex reactions and the variability was slightly better as the mixing operation was improved. The ability to purify and stabilize RNA and to conduct qRT-PCR on-orbit is an important step towards utilizing the ISS as a National Laboratory facility. The ability to get on-orbit data will provide investigators with the opportunity to adjust experimental parameters in real time without the need for sample return and re-flight. The WetLab-2 Project is supported by the Research Integration Office in the ISS Program.
Visual Tracking Using 3D Data and Region-Based Active Contours
2016-09-28
adaptive control strategies which explicitly take uncertainty into account. Filtering methods ranging from the classical Kalman filters valid for...linear systems to the much more general particle filters also fit into this framework in a very natural manner. In particular, the particle filtering ...the number of samples required for accurate filtering increases with the dimension of the system noise. In our approach, we approximate curve
NASA Astrophysics Data System (ADS)
Mukaro, R.; Gasseller, M.; Kufazvinei, C.; Olumekor, L.; Taele, B. M.
2003-08-01
A microcontroller-based multi-sensor temperature measurement and control system that uses a steady-state one-dimensional heat-flow technique for absolute determination of thermal conductivity of a rigid poor conductor using the guarded hot-plate method is described. The objective of this project was to utilize the latest powerful, yet inexpensive, technological developments, sensors, data acquisition and control system, computer and application software, for research and teaching by example. The system uses an ST6220 microcontroller and LM335 temperature sensors for temperature measurement and control. The instrument interfaces to a computer via the serial port using a Turbo C++ programme. LM335Z silicon semiconductor temperature sensors located at different axial locations in the heat source were calibrated and used to measure temperature in the range from room temperature (about 293 K) to 373 K. A zero and span circuit was used in conjunction with an eight-to-one-line data multiplexer to scale the LM335 output signals to fit the 0 5.0 V full-scale input of the microcontroller's on-chip ADC and to sequentially measure temperature at the different locations. Temperature control is achieved by using software-generated pulse-width-modulated signals that control power to the heater. This article emphasizes the apparatus's instrumentation, the computerized data acquisition design, operation and demonstration of the system as a purposeful measurement system that could be easily adopted for use in the undergraduate laboratory. Measurements on a 10 mm thick sample of polyurethane foam at different temperature gradients gave a thermal conductivity of 0.026 +/- 0.004 W m-1 K-1.
Methodology of Global Adult Tobacco Survey (GATS), Malaysia, 2011
Omar, Azahadi; Yusoff, Muhammad Fadhli Mohd; Hiong, Tee Guat; Aris, Tahir; Morton, Jeremy; Pujari, Sameer
2015-01-01
Introduction Malaysia participated in the second phase of the Global Adult Tobacco Survey (GATS) in 2011. GATS, a new component of the Global Tobacco Surveillance System, is a nationally representative household survey of adults 15 years old or above. The objectives of GATS Malaysia were to (i) systematically monitor tobacco use among adults and track key indicators of tobacco control and (ii) track the implementation of some of the Framework Convention of Tobacco Control (FCTC)-recommended demand related policies. Methods GATS Malaysia 2011 was a nationwide cross-sectional survey using multistage stratified sampling to select 5112 nationally representative households. One individual aged 15 years or older was randomly chosen from each selected household and interviewed using handheld device. GATS Core Questionnaire with optional questions was pre-tested and uploaded into handheld devices after repeated quality control processes. Data collectors were trained through a centralized training. Manuals and picture book were prepared to aid in the training of data collectors and during data collection. Field-level data were aggregated on a daily basis and analysed twice a week. Quality controls were instituted to ensure collection of high quality data. Sample weighting and analysis were conducted with the assistance of researchers from the Centers for Disease Control and Prevention, Atlanta, USA Results GATS Malaysia received a total response rate of 85.3% from 5112 adults surveyed. Majority of the respondents were 25–44 years old and Malays. Conclusions The robust methodology used in the GATS Malaysia provides national estimates for tobacco used classified by socio-demographic characteristics and reliable data on various dimensions of tobacco control. PMID:26451348
A robust variable sampling time BLDC motor control design based upon μ-synthesis.
Hung, Chung-Wen; Yen, Jia-Yush
2013-01-01
The variable sampling rate system is encountered in many applications. When the speed information is derived from the position marks along the trajectory, one would have a speed dependent sampling rate system. The conventional fixed or multisampling rate system theory may not work in these cases because the system dynamics include the uncertainties which resulted from the variable sampling rate. This paper derived a convenient expression for the speed dependent sampling rate system. The varying sampling rate effect is then translated into multiplicative uncertainties to the system. The design then uses the popular μ-synthesis process to achieve a robust performance controller design. The implementation on a BLDC motor demonstrates the effectiveness of the design approach.
A Robust Variable Sampling Time BLDC Motor Control Design Based upon μ-Synthesis
Yen, Jia-Yush
2013-01-01
The variable sampling rate system is encountered in many applications. When the speed information is derived from the position marks along the trajectory, one would have a speed dependent sampling rate system. The conventional fixed or multisampling rate system theory may not work in these cases because the system dynamics include the uncertainties which resulted from the variable sampling rate. This paper derived a convenient expression for the speed dependent sampling rate system. The varying sampling rate effect is then translated into multiplicative uncertainties to the system. The design then uses the popular μ-synthesis process to achieve a robust performance controller design. The implementation on a BLDC motor demonstrates the effectiveness of the design approach. PMID:24327804
Gopalakrishnan, V; Subramanian, V; Baskaran, R; Venkatraman, B
2015-07-01
Wireless based custom built aerosol sampling network is designed, developed, and implemented for environmental aerosol sampling. These aerosol sampling systems are used in field measurement campaign, in which sodium aerosol dispersion experiments have been conducted as a part of environmental impact studies related to sodium cooled fast reactor. The sampling network contains 40 aerosol sampling units and each contains custom built sampling head and the wireless control networking designed with Programmable System on Chip (PSoC™) and Xbee Pro RF modules. The base station control is designed using graphical programming language LabView. The sampling network is programmed to operate in a preset time and the running status of the samplers in the network is visualized from the base station. The system is developed in such a way that it can be used for any other environment sampling system deployed in wide area and uneven terrain where manual operation is difficult due to the requirement of simultaneous operation and status logging.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gopalakrishnan, V.; Subramanian, V.; Baskaran, R.
2015-07-15
Wireless based custom built aerosol sampling network is designed, developed, and implemented for environmental aerosol sampling. These aerosol sampling systems are used in field measurement campaign, in which sodium aerosol dispersion experiments have been conducted as a part of environmental impact studies related to sodium cooled fast reactor. The sampling network contains 40 aerosol sampling units and each contains custom built sampling head and the wireless control networking designed with Programmable System on Chip (PSoC™) and Xbee Pro RF modules. The base station control is designed using graphical programming language LabView. The sampling network is programmed to operate in amore » preset time and the running status of the samplers in the network is visualized from the base station. The system is developed in such a way that it can be used for any other environment sampling system deployed in wide area and uneven terrain where manual operation is difficult due to the requirement of simultaneous operation and status logging.« less
ORACLS: A system for linear-quadratic-Gaussian control law design
NASA Technical Reports Server (NTRS)
Armstrong, E. S.
1978-01-01
A modern control theory design package (ORACLS) for constructing controllers and optimal filters for systems modeled by linear time-invariant differential or difference equations is described. Numerical linear-algebra procedures are used to implement the linear-quadratic-Gaussian (LQG) methodology of modern control theory. Algorithms are included for computing eigensystems of real matrices, the relative stability of a matrix, factored forms for nonnegative definite matrices, the solutions and least squares approximations to the solutions of certain linear matrix algebraic equations, the controllability properties of a linear time-invariant system, and the steady state covariance matrix of an open-loop stable system forced by white noise. Subroutines are provided for solving both the continuous and discrete optimal linear regulator problems with noise free measurements and the sampled-data optimal linear regulator problem. For measurement noise, duality theory and the optimal regulator algorithms are used to solve the continuous and discrete Kalman-Bucy filter problems. Subroutines are also included which give control laws causing the output of a system to track the output of a prescribed model.
Xiang, Kun; Li, Yinglei; Ford, William; Land, Walker; Schaffer, J David; Congdon, Robert; Zhang, Jing; Sadik, Omowunmi
2016-02-21
We hereby report the design and implementation of an Autonomous Microbial Cell Culture and Classification (AMC(3)) system for rapid detection of food pathogens. Traditional food testing methods require multistep procedures and long incubation period, and are thus prone to human error. AMC(3) introduces a "one click approach" to the detection and classification of pathogenic bacteria. Once the cultured materials are prepared, all operations are automatic. AMC(3) is an integrated sensor array platform in a microbial fuel cell system composed of a multi-potentiostat, an automated data collection system (Python program, Yocto Maxi-coupler electromechanical relay module) and a powerful classification program. The classification scheme consists of Probabilistic Neural Network (PNN), Support Vector Machines (SVM) and General Regression Neural Network (GRNN) oracle-based system. Differential Pulse Voltammetry (DPV) is performed on standard samples or unknown samples. Then, using preset feature extractions and quality control, accepted data are analyzed by the intelligent classification system. In a typical use, thirty-two extracted features were analyzed to correctly classify the following pathogens: Escherichia coli ATCC#25922, Escherichia coli ATCC#11775, and Staphylococcus epidermidis ATCC#12228. 85.4% accuracy range was recorded for unknown samples, and within a shorter time period than the industry standard of 24 hours.
Improving the Acquisition and Management of Sample Curation Data
NASA Technical Reports Server (NTRS)
Todd, Nancy S.; Evans, Cindy A.; Labasse, Dan
2011-01-01
This paper discusses the current sample documentation processes used during and after a mission, examines the challenges and special considerations needed for designing effective sample curation data systems, and looks at the results of a simulated sample result mission and the lessons learned from this simulation. In addition, it introduces a new data architecture for an integrated sample Curation data system being implemented at the NASA Astromaterials Acquisition and Curation department and discusses how it improves on existing data management systems.
NASA Technical Reports Server (NTRS)
Neiner, G. H.; Cole, G. L.; Arpasi, D. J.
1972-01-01
Digital computer control of a mixed-compression inlet is discussed. The inlet was terminated with a choked orifice at the compressor face station to dynamically simulate a turbojet engine. Inlet diffuser exit airflow disturbances were used. A digital version of a previously tested analog control system was used for both normal shock and restart control. Digital computer algorithms were derived using z-transform and finite difference methods. Using a sample rate of 1000 samples per second, the digital normal shock and restart controls essentially duplicated the inlet analog computer control results. At a sample rate of 100 samples per second, the control system performed adequately but was less stable.
RaPToRS Sample Delivery System
NASA Astrophysics Data System (ADS)
Henchen, Robert; Shibata, Kye; Krieger, Michael; Pogozelski, Edward; Padalino, Stephen; Glebov, Vladimir; Sangster, Craig
2010-11-01
At various labs (NIF, LLE, NRL), activated material samples are used to measure reaction properties. The Rapid Pneumatic Transport of Radioactive Samples (RaPToRS) system quickly and safely moves these radioactive samples through a closed PVC tube via airflow. The carrier travels from the reaction chamber to the control and analysis station, pneumatically braking at the outlet. A reversible multiplexer routes samples from various locations near the shot chamber to the analysis station. Also, the multiplexer allows users to remotely load unactivated samples without manually approaching the reaction chamber. All elements of the system (pneumatic drivers, flow control valves, optical position sensors, multiplexers, Geiger counters, and release gates at the analysis station) can be controlled manually or automatically using a custom LabVIEW interface. A prototype is currently operating at NRL in Washington DC. Prospective facilities for Raptors systems include LLE and NIF.
Backhausen, Lea L.; Herting, Megan M.; Buse, Judith; Roessner, Veit; Smolka, Michael N.; Vetter, Nora C.
2016-01-01
In structural magnetic resonance imaging motion artifacts are common, especially when not scanning healthy young adults. It has been shown that motion affects the analysis with automated image-processing techniques (e.g., FreeSurfer). This can bias results. Several developmental and adult studies have found reduced volume and thickness of gray matter due to motion artifacts. Thus, quality control is necessary in order to ensure an acceptable level of quality and to define exclusion criteria of images (i.e., determine participants with most severe artifacts). However, information about the quality control workflow and image exclusion procedure is largely lacking in the current literature and the existing rating systems differ. Here, we propose a stringent workflow of quality control steps during and after acquisition of T1-weighted images, which enables researchers dealing with populations that are typically affected by motion artifacts to enhance data quality and maximize sample sizes. As an underlying aim we established a thorough quality control rating system for T1-weighted images and applied it to the analysis of developmental clinical data using the automated processing pipeline FreeSurfer. This hands-on workflow and quality control rating system will aid researchers in minimizing motion artifacts in the final data set, and therefore enhance the quality of structural magnetic resonance imaging studies. PMID:27999528
High-throughput biological small-angle X-ray scattering with a robotically loaded capillary cell
Nielsen, S. S.; Møller, M.; Gillilan, R. E.
2012-01-01
With the rise in popularity of biological small-angle X-ray scattering (BioSAXS) measurements, synchrotron beamlines are confronted with an ever-increasing number of samples from a wide range of solution conditions. To meet these demands, an increasing number of beamlines worldwide have begun to provide automated liquid-handling systems for sample loading. This article presents an automated sample-loading system for BioSAXS beamlines, which combines single-channel disposable-tip pipetting with a vacuum-enclosed temperature-controlled capillary flow cell. The design incorporates an easily changeable capillary to reduce the incidence of X-ray window fouling and cross contamination. Both the robot-control and the data-processing systems are written in Python. The data-processing code, RAW, has been enhanced with several new features to form a user-friendly BioSAXS pipeline for the robot. The flow cell also supports efficient manual loading and sample recovery. An effective rinse protocol for the sample cell is developed and tested. Fluid dynamics within the sample capillary reveals a vortex ring pattern of circulation that redistributes radiation-damaged material. Radiation damage is most severe in the boundary layer near the capillary surface. At typical flow speeds, capillaries below 2 mm in diameter are beginning to enter the Stokes (creeping flow) regime in which mixing due to oscillation is limited. Analysis within this regime shows that single-pass exposure and multiple-pass exposure of a sample plug are functionally the same with regard to exposed volume when plug motion reversal is slow. The robot was tested on three different beamlines at the Cornell High-Energy Synchrotron Source, with a variety of detectors and beam characteristics, and it has been used successfully in several published studies as well as in two introductory short courses on basic BioSAXS methods. PMID:22509071
Power system distributed oscilation detection based on Synchrophasor data
NASA Astrophysics Data System (ADS)
Ning, Jiawei
Along with increasing demand for electricity, integration of renewable energy and deregulation of power market, power industry is facing unprecedented challenges nowadays. Within the last couple of decades, several serious blackouts have been taking place in United States. As an effective approach to prevent that, power system small signal stability monitoring has been drawing more interests and attentions from researchers. With wide-spread implementation of Synchrophasors around the world in the last decade, power systems real-time online monitoring becomes much more feasible. Comparing with planning study analysis, real-time online monitoring would benefit control room operators immediately and directly. Among all online monitoring methods, Oscillation Modal Analysis (OMA), a modal identification method based on routine measurement data where the input is unmeasured ambient excitation, is a great tool to evaluate and monitor power system small signal stability. Indeed, high sampling Synchrophasor data around power system is fitted perfectly as inputs to OMA. Existing methods in OMA for power systems are all based on centralized algorithms applying at control centers only; however, with rapid growing number of online Synchrophasors the computation burden at control centers is and will be continually exponentially expanded. The increasing computation time at control center compromises the real-time feature of online monitoring. The communication efforts between substation and control center will also be out of reach. Meanwhile, it is difficult or even impossible for centralized algorithms to detect some poorly damped local modes. In order to avert previous shortcomings of centralized OMA methods and embrace the new changes in the power systems, two new distributed oscillation detection methods with two new decentralized structures are presented in this dissertation. Since the new schemes brought substations into the big oscillation detection picture, the proposed methods could achieve faster and more reliable results. Subsequently, this claim is tested and approved by test results of IEEE Two-area simulation test system and real power system historian synchrophasor data case studies.
An Experimental Characterization System for Deep Ultra-Violet (UV) Photoresists
NASA Astrophysics Data System (ADS)
Drako, Dean M.; Partlo, William N.; Oldham, William G.; Neureuther, Andrew R.
1989-08-01
A versatile system designed specifically for experimental automated photoresist characterization has been constructed utilizing an excimer laser source for exposure at 248nm. The system was assembled, as much as possible, from commercially available components in order to facilitate its replication. The software and hardware are completely documented in a University of California-Berkeley Engineering Research Lab Memo. An IBM PC-AT compatible computer controls an excimer laser, operates a Fourier Transform Infrared (FTIR) Spectrometer, measures and records the energy of each laser pulse (incident, reflected, and transmitted), opens and closes shutters, and operates two linear stages for sample movement. All operations (except FTIR data reduction) are managed by a control program written in the "C" language. The system is capable of measuring total exposure dose, performing bleaching measurements, creating and recording exposure pulse sequences, and generating exposure patterns suitable for multiple channel monitoring of the development. The total exposure energy, energy per pulse, and pulse rate are selectable over a wide range. The system contains an in-situ Fourier Transform Infrared Spectrometer for qualitative and quantitative analysis of the photoresist baking and exposure processes (baking is not done in-situ). FIIR may be performed in transmission or reflection. The FTIR data will form the basis of comprehensive multi-state resist models. The system's versatility facilitates the development of new automated and repeatable experiments. Simple controlling software, utilizing the provided interface sub-routines, can be written to control new experiments and collect data.
SCADA data and the quantification of hazardous events for QMRA.
Nilsson, P; Roser, D; Thorwaldsdotter, R; Petterson, S; Davies, C; Signor, R; Bergstedt, O; Ashbolt, N
2007-01-01
The objective of this study was to assess the use of on-line monitoring to support the QMRA at water treatment plants studied in the EU MicroRisk project. SCADA data were obtained from three Catchment-to-Tap Systems (CTS) along with system descriptions, diary records, grab sample data and deviation reports. Particular attention was paid to estimating hazardous event frequency, duration and magnitude. Using Shewart and CUSUM we identified 'change-points' corresponding to events of between 10 min and >1 month duration in timeseries data. Our analysis confirmed it is possible to quantify hazardous event durations from turbidity, chlorine residual and pH records and distinguish them from non-hazardous variability in the timeseries dataset. The durations of most 'events' were short-term (0.5-2.3 h). These data were combined with QMRA to estimate pathogen infection risk arising from such events as chlorination failure. While analysis of SCADA data alone could identify events provisionally, its interpretation was severely constrained in the absence of diary records and other system information. SCADA data analysis should only complement traditional water sampling, rather than replace it. More work on on-line data management, quality control and interpretation is needed before it can be used routinely for event characterization.
Van der Merwe, Deon; Price, Kevin P
2015-03-27
Harmful algal blooms (HABs) degrade water quality and produce toxins. The spatial distribution of HAbs may change rapidly due to variations wind, water currents, and population dynamics. Risk assessments, based on traditional sampling methods, are hampered by the sparseness of water sample data points, and delays between sampling and the availability of results. There is a need for local risk assessment and risk management at the spatial and temporal resolution relevant to local human and animal interactions at specific sites and times. Small, unmanned aircraft systems can gather color-infrared reflectance data at appropriate spatial and temporal resolutions, with full control over data collection timing, and short intervals between data gathering and result availability. Data can be interpreted qualitatively, or by generating a blue normalized difference vegetation index (BNDVI) that is correlated with cyanobacterial biomass densities at the water surface, as estimated using a buoyant packed cell volume (BPCV). Correlations between BNDVI and BPCV follow a logarithmic model, with r(2)-values under field conditions from 0.77 to 0.87. These methods provide valuable information that is complimentary to risk assessment data derived from traditional risk assessment methods, and could help to improve risk management at the local level.
Van der Merwe, Deon; Price, Kevin P.
2015-01-01
Harmful algal blooms (HABs) degrade water quality and produce toxins. The spatial distribution of HAbs may change rapidly due to variations wind, water currents, and population dynamics. Risk assessments, based on traditional sampling methods, are hampered by the sparseness of water sample data points, and delays between sampling and the availability of results. There is a need for local risk assessment and risk management at the spatial and temporal resolution relevant to local human and animal interactions at specific sites and times. Small, unmanned aircraft systems can gather color-infrared reflectance data at appropriate spatial and temporal resolutions, with full control over data collection timing, and short intervals between data gathering and result availability. Data can be interpreted qualitatively, or by generating a blue normalized difference vegetation index (BNDVI) that is correlated with cyanobacterial biomass densities at the water surface, as estimated using a buoyant packed cell volume (BPCV). Correlations between BNDVI and BPCV follow a logarithmic model, with r2-values under field conditions from 0.77 to 0.87. These methods provide valuable information that is complimentary to risk assessment data derived from traditional risk assessment methods, and could help to improve risk management at the local level. PMID:25826055
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lanekoff, Ingela T.; Heath, Brandi S.; Liyu, Andrey V.
2012-10-02
An automated platform has been developed for acquisition and visualization of mass spectrometry imaging (MSI) data using nanospray desorption electrospray ionization (nano-DESI). The new system enables robust operation of the nano-DESI imaging source over many hours. This is achieved by controlling the distance between the sample and the probe by mounting the sample holder onto an automated XYZ stage and defining the tilt of the sample plane. This approach is useful for imaging of relatively flat samples such as thin tissue sections. Custom software called MSI QuickView was developed for visualization of large data sets generated in imaging experiments. MSImore » QuickView enables fast visualization of the imaging data during data acquisition and detailed processing after the entire image is acquired. The performance of the system is demonstrated by imaging rat brain tissue sections. High resolution mass analysis combined with MS/MS experiments enabled identification of lipids and metabolites in the tissue section. In addition, high dynamic range and sensitivity of the technique allowed us to generate ion images of low-abundance isobaric lipids. High-spatial resolution image acquired over a small region of the tissue section revealed the spatial distribution of an abundant brain metabolite, creatine, in the white and gray matter that is consistent with the literature data obtained using magnetic resonance spectroscopy.« less
Software framework for the upcoming MMT Observatory primary mirror re-aluminization
NASA Astrophysics Data System (ADS)
Gibson, J. Duane; Clark, Dusty; Porter, Dallan
2014-07-01
Details of the software framework for the upcoming in-situ re-aluminization of the 6.5m MMT Observatory (MMTO) primary mirror are presented. This framework includes: 1) a centralized key-value store and data structure server for data exchange between software modules, 2) a newly developed hardware-software interface for faster data sampling and better hardware control, 3) automated control algorithms that are based upon empirical testing, modeling, and simulation of the aluminization process, 4) re-engineered graphical user interfaces (GUI's) that use state-of-the-art web technologies, and 5) redundant relational databases for data logging. Redesign of the software framework has several objectives: 1) automated process control to provide more consistent and uniform mirror coatings, 2) optional manual control of the aluminization process, 3) modular design to allow flexibility in process control and software implementation, 4) faster data sampling and logging rates to better characterize the approximately 100-second aluminization event, and 5) synchronized "real-time" web application GUI's to provide all users with exactly the same data. The framework has been implemented as four modules interconnected by a data store/server. The four modules are integrated into two Linux system services that start automatically at boot-time and remain running at all times. Performance of the software framework is assessed through extensive testing within 2.0 meter and smaller coating chambers at the Sunnyside Test Facility. The redesigned software framework helps ensure that a better performing and longer lasting coating will be achieved during the re-aluminization of the MMTO primary mirror.
Zagar, Agata Karolina; Zagar, Robert John; Bartikowski, Boris; Busch, Kenneth G
2009-02-01
Data from youth studied by Zagar and colleagues were randomly sampled to create groups of controls and abused, delinquent, violent, and homicidal youth (n=30 in each). Estimated costs of raising a nondelinquent youth from birth to 17 yr. were compared with the average costs incurred by other youth in each group. Estimates of living expenses, direct and indirect costs of victimization, and criminal justice system expenditures were summed. Groups differed significantly on total expenses, victimization costs, and criminal justice expenditures. Mean total costs for a homicidal youth were estimated at $3,935,433, while those for a control youth were $150,754. Abused, delinquent, and violent youth had average total expenses roughly double the total mean costs of controls. Prevention of dropout, alcoholism, addiction, career delinquency, or homicide justifies interception and empirical treatment on a cost-benefit basis, but also based on the severe personal costs to the victims and to the youth themselves.
MISSE in the Materials and Processes Technical Information System (MAPTIS )
NASA Technical Reports Server (NTRS)
Burns, DeWitt; Finckenor, Miria; Henrie, Ben
2013-01-01
Materials International Space Station Experiment (MISSE) data is now being collected and distributed through the Materials and Processes Technical Information System (MAPTIS) at Marshall Space Flight Center in Huntsville, Alabama. MISSE data has been instrumental in many programs and continues to be an important source of data for the space community. To facilitate great access to the MISSE data the International Space Station (ISS) program office and MAPTIS are working to gather this data into a central location. The MISSE database contains information about materials, samples, and flights along with pictures, pdfs, excel files, word documents, and other files types. Major capabilities of the system are: access control, browsing, searching, reports, and record comparison. The search capabilities will search within any searchable files so even if the desired meta-data has not been associated data can still be retrieved. Other functionality will continue to be added to the MISSE database as the Athena Platform is expanded
NASA Astrophysics Data System (ADS)
Gaffney, Monique Suzanne
1998-11-01
Metalorganic chemical vapor deposition (MOCVD) is a process used to manufacture electronic and optoelectronic devices that has traditionally lacked real-time growth monitoring and control. Controlling the growth rate and composition using the existing sensors, as well as advanced monitoring systems developed in-house, is shown to improve device quality. Specific MOCVD growth objectives are transformed into controller performance goals. Group III bubbler concentration variations, which perturb both growth rate and composition precision, are identified to be the primary disturbances. First a feed forward control system was investigated, which used an ultrasonic concentration monitor, located upstream in the process. This control strategy resulted in improved regulation of the gallium delivery rate by cancelling the sensed gallium bubbler concentration disturbances via the injection mass flow controller. The controller performance is investigated by growing GaInAs/InP superlattices. Results of growths performed under normal operating conditions and also under large perturbations include X-ray diffraction from the samples as well as real-time sensor signal data. High quality superlattices that display up to eight orders of satellite peaks are obtained under the feed forward compensation scheme, demonstrating improved layer-to-layer reproducibility of thickness and composition. The success of the feed forward control demonstration led to the development of a more complex downstream feedback control system. An ultraviolet absorption monitor was fabricated and retrofitted as a feedback control signal. A control-oriented model of the downstream process was developed for the feedback controller synthesis. Although challenged with both the photolysis and multi-gas detection issues common to UV absorption monitors, closed loop control with the UV sensor was performed and proved to be an effective method of disturbance rejection. An InP/GaInAs test structure was grown under both open and closed loop conditions. During the growth of a bulk GaInAs layer, an indium concentration disturbance was injected by way of the bubbler pressure control valve. The controller goal was to reject this concentration disturbance. The UV absorption real-time data, as well as both X-ray diffraction and photoluminescence post-growth sample measurements were used to evaluate the controller performance. All results indicate that the closed loop control system greatly improved the quality of the perturbed growth.
Lu, Zhan; Zhang, Jianyi; Xu, Lizhou; Li, Yanbin; Chen, Siyu; Ye, Zunzhong; Wang, Jianping
2017-01-01
A simple, highly-automated instrument system used for on-site detection of foodborne pathogens based on fluorescence was designed, fabricated, and preliminarily tested in this paper. A corresponding method has been proved effective in our previous studies. This system utilizes a light-emitting diode (LED) to excite fluorescent labels and a spectrometer to record the fluorescence signal from samples. A rotation stage for positioning and switching samples was innovatively designed for high-throughput detection, ten at most in one single run. We also developed software based on LabVIEW for data receiving, processing, and the control of the whole system. In the test of using a pure quantum dot (QD) solution as a standard sample, detection results from this home-made system were highly-relevant with that from a well-commercialized product and even slightly better reproducibility was found. And in the test of three typical kinds of food-borne pathogens, fluorescence signals recorded by this system are highly proportional to the variation of the sample concentration, with a satisfied limit of detection (LOD) (nearly 102–103 CFU·mL−1 in food samples). Additionally, this instrument system is low-cost and easy-to-use, showing a promising potential for on-site rapid detection of food-borne pathogens. PMID:28241478
Lu, Zhan; Zhang, Jianyi; Xu, Lizhou; Li, Yanbin; Chen, Siyu; Ye, Zunzhong; Wang, Jianping
2017-02-23
A simple, highly-automated instrument system used for on-site detection of foodborne pathogens based on fluorescence was designed, fabricated, and preliminarily tested in this paper. A corresponding method has been proved effective in our previous studies. This system utilizes a light-emitting diode (LED) to excite fluorescent labels and a spectrometer to record the fluorescence signal from samples. A rotation stage for positioning and switching samples was innovatively designed for high-throughput detection, ten at most in one single run. We also developed software based on LabVIEW for data receiving, processing, and the control of the whole system. In the test of using a pure quantum dot (QD) solution as a standard sample, detection results from this home-made system were highly-relevant with that from a well-commercialized product and even slightly better reproducibility was found. And in the test of three typical kinds of food-borne pathogens, fluorescence signals recorded by this system are highly proportional to the variation of the sample concentration, with a satisfied limit of detection (LOD) (nearly 10²-10³ CFU·mL -1 in food samples). Additionally, this instrument system is low-cost and easy-to-use, showing a promising potential for on-site rapid detection of food-borne pathogens.
An adaptive management approach to controlling suburban deer
Nielson, C.K.; Porter, W.F.; Underwood, H.B.
1997-01-01
Distance sight-resight sampling has particular relevance to aerial surveys, in which height above ground and aircraft speed make the critical assumption of certain detection on the track-line unrealistic. Recent developments in distance sight-resight theory have left practical issues related to data collection as the major impediment to widespread use of distance sight-resight sampling in aerial surveys. We describe and evaluate a system to automatically log, store, and process data from distance sight-resight aerial surveys. The system has a primary digital system and a secondary audio system. The digital system comprises a sighting 'gun' and small keypad for each observer, a global positioning system (GPS) receiver, and an altimeter interface, all linked to a central laptop computer. The gun is used to record time and angle of declination from the horizon of sighted groups of animals as they pass the aircraft. The keypad is used to record information on species and group size. The altimeter interface records altitude from the aircraft's radar altimeter, and the GPS receiver provides location data at user-definable intervals. We wrote software to import data into a database and convert it into a form appropriate for distance sight-resight analyses. Perpendicular distance of sighted groups of animals from the flight path is calculated from altitude and angle of declination. Time, angle of declination, species, and group size of sightings by independent observers on the same side of the aircraft are used as criteria to classify single and duplicate sightings, allowing testing of the critical distance sampling assumption (g(0)=1) and estimation of g(0) if that assumption fails. An audio system comprising headphones for each observer and a 4-track tape recorder allows recording of data that are difficult to accommodate in the digital system and provides a backup to the digital system. We evaluated the system by conducting experimental surveys and reviewing results from actual surveys.
Data Quality Control of the French Permanent Broadband Network in the RESIF Framework
NASA Astrophysics Data System (ADS)
Grunberg, Marc; Lambotte, Sophie; Engels, Fabien; Dretzen, Remi; Hernandez, Alain
2014-05-01
In the framework of the RESIF (Réseau Sismologique et géodésique Français) project, a new information system is being setting up, allowing the improvement of the management and the distribution of high quality data from the different elements of RESIF and the associated networks. Within this information system, EOST (in Strasbourg) is in charge of collecting real-time permanent broadband seismic waveform, and performing Quality Control on these data. The real-time and validated data set are pushed to the French National Distribution Center (Isterre/Grenoble) in order to make them publicly available. Furthermore EOST hosts the BCSF-ReNaSS, in charge of the French metropolitan seismic bulletin. This allows to benefit from some high-end quality control based on the national and world-wide seismicity. Here we present first the real-time seismic data flow from the stations of the French National Broad Band Network to EOST, and then, the data Quality Control procedures that were recently installed, including some new developments. The data Quality Control consists in applying a variety of subprocesses to check the consistency of the whole system and process from the stations to the data center. This allows us to verify that instruments and data transmission are operating correctly. Moreover analysis of the ambient noise helps to characterize intrinsic seismic quality of the stations and to identify other kind of disturbances. The deployed Quality Control consist in a pipeline that starts with low-level procedures : check the real-time miniseed data file (file naming convention, data integrity), check for inconsistencies between waveform and meta-data (channel name, sample rate, etc.), compute waveform statistics (data availability, gap/overlap, mean, rms, time quality, spike). It is followed by some high-level procedures such as : power spectral density computation (PSD), STA/LTA computation to be correlated to the seismicity, phases picking and stations magnitudes discrepancies. The results of quality control is visualized through a web interface. This latter gathers data from different information systems to provide a global view on last events that could impact the data (like intervention on site or seismic events, etc.). This work is still an ongoing project. We intend to add more sophisticated procedures to enhanced our data Quality Control. Among them, we will deploy a seismic moment tensor inversion tool for amplitude, time and polarity control and a noise correlation procedure for time drift detections.
Design, implementation and flight testing of PIF autopilots for general aviation aircraft
NASA Technical Reports Server (NTRS)
Broussard, J. R.
1983-01-01
The designs of Proportional-Integrated-Filter (PIF) auto-pilots for a General Aviation (NAVION) aircraft are presented. The PIF autopilot uses the sampled-data regulator and command generator tracking to determine roll select, pitch select, heading select, altitude select and localizer/glideslope capture and hold autopilot modes. The PIF control law uses typical General Aviation sensors for state feedback, command error integration for command tracking, digital complementary filtering and analog prefiltering for sensor noise suppression, a control filter for computation delay accommodation and the incremental form to eliminate trim values in implementation. Theoretical developments described in detail, were needed to combine the sampled-data regulator with command generator tracking for use as a digital flight control system. The digital PIF autopilots are evaluated using closed-loop eigenvalues and linear simulations. The implementation of the PIF autopilots in a digital flight computer using a high order language (FORTRAN) is briefly described. The successful flight test results for each PIF autopilot mode is presented.
40 CFR 86.605-88 - Maintenance of records; submittal of information.
Code of Federal Regulations, 2011 CFR
2011-07-01
..., pressure increase across the pump, and the temperature set point of the temperature control system. (2... samples are being collected. (3) Humidity of dilution air. (4) Manufacturer, model, type and serial number..., ambient temperature and humidity. (2) Data and time of day. (ii) In lieu of recording test equipment...
40 CFR 86.605-88 - Maintenance of records; submittal of information.
Code of Federal Regulations, 2012 CFR
2012-07-01
..., pressure increase across the pump, and the temperature set point of the temperature control system. (2... samples are being collected. (3) Humidity of dilution air. (4) Manufacturer, model, type and serial number..., ambient temperature and humidity. (2) Data and time of day. (ii) In lieu of recording test equipment...
40 CFR 86.605-88 - Maintenance of records; submittal of information.
Code of Federal Regulations, 2013 CFR
2013-07-01
..., pressure increase across the pump, and the temperature set point of the temperature control system. (2... samples are being collected. (3) Humidity of dilution air. (4) Manufacturer, model, type and serial number..., ambient temperature and humidity. (2) Data and time of day. (ii) In lieu of recording test equipment...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Robinson Khosah
2007-07-31
Advanced Technology Systems, Inc. (ATS) was contracted by the U. S. Department of Energy's National Energy Technology Laboratory (DOE-NETL) to develop a state-of-the-art, scalable and robust web-accessible database application to manage the extensive data sets resulting from the DOE-NETL-sponsored ambient air monitoring programs in the upper Ohio River valley region. The data management system was designed to include a web-based user interface that will allow easy access to the data by the scientific community, policy- and decision-makers, and other interested stakeholders, while providing detailed information on sampling, analytical and quality control parameters. In addition, the system will provide graphical analyticalmore » tools for displaying, analyzing and interpreting the air quality data. The system will also provide multiple report generation capabilities and easy-to-understand visualization formats that can be utilized by the media and public outreach/educational institutions. The project was conducted in two phases. Phase One included the following tasks: (1) data inventory/benchmarking, including the establishment of an external stakeholder group; (2) development of a data management system; (3) population of the database; (4) development of a web-based data retrieval system, and (5) establishment of an internal quality assurance/quality control system on data management. Phase Two involved the development of a platform for on-line data analysis. Phase Two included the following tasks: (1) development of a sponsor and stakeholder/user website with extensive online analytical tools; (2) development of a public website; (3) incorporation of an extensive online help system into each website; and (4) incorporation of a graphical representation (mapping) system into each website. The project is now technically completed.« less
Demodulator for carrier transducers
NASA Technical Reports Server (NTRS)
Roller, R. F. (Inventor)
1974-01-01
A carrier type transducer is supplied with a carrier wave via an audio amplifier, a filter, a frequency divider, and an oscillator. The carrier is modulated in accordance with the parameter being measured by the transducer and is fed to the input of a digital data system which may include a voltmeter. The output of the oscillator and the output of each stage of the divider are fed to an AND or a NAND gate and suitable variable and fixed delay circuits to the command input of the digital data system. With this arrangement, the digital data system is commanded to sample at the proper time so that the average voltage of the modulated carrier is measured. It may be utilized with ancillary circuitry for control of the parameter
An investigation of potential applications of OP-SAPS: Operational sampled analog processors
NASA Technical Reports Server (NTRS)
Parrish, E. A.; Mcvey, E. S.
1976-01-01
The impact of charge-coupled device (CCD) processors on future instrumentation was investigated. The CCD devices studied process sampled analog data and are referred to as OP-SAPS - operational sampled analog processors. Preliminary studies into various architectural configurations for systems composed of OP-SAPS show that they have potential in such diverse applications as pattern recognition and automatic control. It appears probable that OP-SAPS may be used to construct computing structures which can serve as special peripherals to large-scale computer complexes used in real time flight simulation. The research was limited to the following benchmark programs: (1) face recognition, (2) voice command and control, (3) terrain classification, and (4) terrain identification. A small amount of effort was spent on examining a method by which OP-SAPS may be used to decrease the limiting ground sampling distance encountered in remote sensing from satellites.
Mansano, Raul K.; Godoy, Eduardo P.; Porto, Arthur J. V.
2014-01-01
Recent advances in wireless networking technology and the proliferation of industrial wireless sensors have led to an increasing interest in using wireless networks for closed loop control. The main advantages of Wireless Networked Control Systems (WNCSs) are the reconfigurability, easy commissioning and the possibility of installation in places where cabling is impossible. Despite these advantages, there are two main problems which must be considered for practical implementations of WNCSs. One problem is the sampling period constraint of industrial wireless sensors. This problem is related to the energy cost of the wireless transmission, since the power supply is limited, which precludes the use of these sensors in several closed-loop controls. The other technological concern in WNCS is the energy efficiency of the devices. As the sensors are powered by batteries, the lowest possible consumption is required to extend battery lifetime. As a result, there is a compromise between the sensor sampling period, the sensor battery lifetime and the required control performance for the WNCS. This paper develops a model-based soft sensor to overcome these problems and enable practical implementations of WNCSs. The goal of the soft sensor is generating virtual data allowing an actuation on the process faster than the maximum sampling period available for the wireless sensor. Experimental results have shown the soft sensor is a solution to the sampling period constraint problem of wireless sensors in control applications, enabling the application of industrial wireless sensors in WNCSs. Additionally, our results demonstrated the soft sensor potential for implementing energy efficient WNCS through the battery saving of industrial wireless sensors. PMID:25529208
1991-07-01
Systems Audits ; Preventive Maintenance; Data Assessment Procedures; Corrective Action; Quality Assurance Reports; and Site Management. General Data Quality...for verification and quality control audits . A copy of the QAPP will be in the possession of field sampling teams for all sampling efforts...L cc- acca C wi w-C P-3-- C I- usu 2; - at C4~~2 Q (( c ccc c-c C- 0L U -C 3- au S - - C3 . - acu tr - -a CL.43 C LU -C2 -C ul 0it S w- -C u Ocw - z i
Cuffney, T.F.; Gurtz, M.E.; Meador, M.R.
1993-01-01
Benthic invertebrate samples are collected as part of the U.S. Geological Survey's National Water-Quality Assessment Program. This is a perennial, multidisciplinary program that integrates biological, physical, and chemical indicators of water quality to evaluate status and trends and to develop an understanding of the factors controlling observed water quality. The Program examines water quality in 60 study units (coupled ground- and surface-water systems) that encompass most of the conterminous United States and parts of Alaska and Hawaii. Study-unit teams collect and process qualitative and semi-quantitative invertebrate samples according to standardized procedures. These samples are processed (elutriated and subsampled) in the field to produce as many as four sample components: large-rare, main-body, elutriate, and split. Each sample component is preserved in 10-percent formalin, and two components, large-rare and main-body, are sent to contract laboratories for further processing. The large-rare component is composed of large invertebrates that are removed from the sample matrix during field processing and placed in one or more containers. The main-body sample component consists of the remaining sample materials (sediment, detritus, and invertebrates) and is subsampled in the field to achieve a volume of 750 milliliters or less. The remaining two sample components, elutriate and split, are used for quality-assurance and quality-control purposes. Contract laboratories are used to identify and quantify invertebrates from the large-rare and main-body sample components according to the procedures and guidelines specified within this document. These guidelines allow the use of subsampling techniques to reduce the volume of sample material processed and to facilitate identifications. These processing procedures and techniques may be modified if the modifications provide equal or greater levels of accuracy and precision. The intent of sample processing is to determine the quantity of each taxon present in the semi-quantitative samples or to list the taxa present in qualitative samples. The processing guidelines provide standardized laboratory forms, sample labels, detailed sample processing flow charts, standardized format for electronic data, quality-assurance procedures and checks, sample tracking standards, and target levels for taxonomic determinations. The contract laboratory (1) is responsible for identifications and quantifications, (2) constructs reference collections, (3) provides data in hard copy and electronic forms, (4) follows specified quality-assurance and quality-control procedures, and (5) returns all processed and unprocessed portions of the samples. The U.S. Geological Survey's Quality Management Group maintains a Biological Quality-Assurance Unit, located at the National Water-Quality Laboratory, Arvada, Colorado, to oversee the use of contract laboratories and ensure the quality of data obtained from these laboratories according to the guidelines established in this document. This unit establishes contract specifications, reviews contractor performance (timeliness, accuracy, and consistency), enters data into the National Water Information System-II data base, maintains in-house reference collections, deposits voucher specimens in outside museums, and interacts with taxonomic experts within and outside the U.S. Geological Survey. This unit also modifies the existing sample processing and quality-assurance guidelines, establishes criteria and testing procedures for qualifying potential contract laboratories, identifies qualified taxonomic experts, and establishes voucher collections.
Extending and expanding the life of older current meters
Strahle, W.J.; Martini, Marinna A.
1995-01-01
The EG&G Model 610 VACM and Model 630 VMCM are standards for ocean current measurements. It is simple to add peripheral sensors to the data stream of the VACM by use of add-on CMOS circuitry. The firmware control of the VMCM makes it virtually impossible to add sampling of additional sensors. Most of the electronic components used in the VACM are obsolete or difficult to replace and the VMCM will soon follow suit. As a result, the USGS joined WHOI in the development of a PCMCIA data storage system to replace the cassette recording system in the VACM. Using the same PCMCIA recording package as the controller and recorder for the VMCM, a user-friendly VMCM is being designed. PCMCIA cards are rapidly becoming an industry standard with a wide range of storage capacities. By upgrading the VACM and VMCM to PCMCIA storage systems with a flexible microprocessor, they will continue to be viable instruments.
Marshall, Charla; Sturk-Andreaggi, Kimberly; Daniels-Higginbotham, Jennifer; Oliver, Robert Sean; Barritt-Ross, Suzanne; McMahon, Timothy P
2017-11-01
Next-generation ancient DNA technologies have the potential to assist in the analysis of degraded DNA extracted from forensic specimens. Mitochondrial genome (mitogenome) sequencing, specifically, may be of benefit to samples that fail to yield forensically relevant genetic information using conventional PCR-based techniques. This report summarizes the Armed Forces Medical Examiner System's Armed Forces DNA Identification Laboratory's (AFMES-AFDIL) performance evaluation of a Next-Generation Sequencing protocol for degraded and chemically treated past accounting samples. The procedure involves hybridization capture for targeted enrichment of mitochondrial DNA, massively parallel sequencing using Illumina chemistry, and an automated bioinformatic pipeline for forensic mtDNA profile generation. A total of 22 non-probative samples and associated controls were processed in the present study, spanning a range of DNA quantity and quality. Data were generated from over 100 DNA libraries by ten DNA analysts over the course of five months. The results show that the mitogenome sequencing procedure is reliable and robust, sensitive to low template (one ng control DNA) as well as degraded DNA, and specific to the analysis of the human mitogenome. Haplotypes were overall concordant between NGS replicates and with previously generated Sanger control region data. Due to the inherent risk for contamination when working with low-template, degraded DNA, a contamination assessment was performed. The consumables were shown to be void of human DNA contaminants and suitable for forensic use. Reagent blanks and negative controls were analyzed to determine the background signal of the procedure. This background signal was then used to set analytical and reporting thresholds, which were designated at 4.0X (limit of detection) and 10.0X (limit of quantiation) average coverage across the mitogenome, respectively. Nearly all human samples exceeded the reporting threshold, although coverage was reduced in chemically treated samples resulting in a ∼58% passing rate for these poor-quality samples. A concordance assessment demonstrated the reliability of the NGS data when compared to known Sanger profiles. One case sample was shown to be mixed with a co-processed sample and two reagent blanks indicated the presence of DNA above the analytical threshold. This contamination was attributed to sequencing crosstalk from simultaneously sequenced high-quality samples to include the positive control. Overall this study demonstrated that hybridization capture and Illumina sequencing provide a viable method for mitogenome sequencing of degraded and chemically treated skeletal DNA samples, yet may require alternative measures of quality control. Copyright © 2017 The Authors. Published by Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Hsu, L.; Lehnert, K. A.; Carbotte, S. M.; Arko, R. A.; Ferrini, V.; O'hara, S. H.; Walker, J. D.
2012-12-01
The Integrated Earth Data Applications (IEDA) facility maintains multiple data systems with a wide range of solid earth data types from the marine, terrestrial, and polar environments. Examples of the different data types include syntheses of ultra-high resolution seafloor bathymetry collected on large collaborative cruises and analytical geochemistry measurements collected by single investigators in small, unique projects. These different data types have historically been channeled into separate, discipline-specific databases with search and retrieval tailored for the specific data type. However, a current major goal is to integrate data from different systems to allow interdisciplinary data discovery and scientific analysis. To increase discovery and access across these heterogeneous systems, IEDA employs several unique IDs, including sample IDs (International Geo Sample Number, IGSN), person IDs (GeoPass ID), funding award IDs (NSF Award Number), cruise IDs (from the Marine Geoscience Data System Expedition Metadata Catalog), dataset IDs (DOIs), and publication IDs (DOIs). These IDs allow linking of a sample registry (System for Earth SAmple Registration), data libraries and repositories (e.g. Geochemical Research Library, Marine Geoscience Data System), integrated synthesis databases (e.g. EarthChem Portal, PetDB), and investigator services (IEDA Data Compliance Tool). The linked systems allow efficient discovery of related data across different levels of granularity. In addition, IEDA data systems maintain links with several external data systems, including digital journal publishers. Links have been established between the EarthChem Portal and ScienceDirect through publication DOIs, returning sample-level objects and geochemical analyses for a particular publication. Linking IEDA-hosted data to digital publications with IGSNs at the sample level and with IEDA-allocated dataset DOIs are under development. As an example, an individual investigator could sign up for a GeoPass account ID, write a proposal to NSF and create a data plan using the IEDA Data Management Plan Tool. Having received the grant, the investigator then collects rock samples on a scientific cruise from dredges and registers the samples with IGSNs. The investigator then performs analytical geochemistry on the samples, and submits the full dataset to the Geochemical Resource Library for a dataset DOI. Finally, the investigator writes an article that is published in Science Direct. Knowing any of the following IDs: Investigator GeoPass ID, NSF Award Number, Cruise ID, Sample IGSNs, dataset DOI, or publication DOI, a user would be able to navigate to all samples, datasets, and publications in IEDA and external systems. Use of persistent identifiers to link heterogeneous data systems in IEDA thus increases access, discovery, and proper citation of hard-earned investigator datasets.
Feedback quantum control of molecular electronic population transfer
NASA Astrophysics Data System (ADS)
Bardeen, Christopher J.; Yakovlev, Vladislav V.; Wilson, Kent R.; Carpenter, Scott D.; Weber, Peter M.; Warren, Warren S.
1997-11-01
Feedback quantum control, where the sample `teaches' a computer-controlled arbitrary lightform generator to find the optimal light field, is experimentally demonstrated for a molecular system. Femtosecond pulses tailored by a computer-controlled acousto-optic pulse shaper excite fluorescence from laser dye molecules in solution. Fluorescence and laser power are monitored, and the computer uses the experimental data and a genetic algorithm to optimize population transfer from ground to first excited state. Both efficiency (the ratio of excited state population to laser energy) and effectiveness (total excited state population) are optimized. Potential use as an `automated theory tester' is discussed.
Huffaker, S Kirk; Safavi, Kamran; Spangberg, Larz S W; Kaufman, Blythe
2010-08-01
The present investigation evaluated the ability of a new passive sonic irrigation (sonic group) system (EndoActivator) to eliminate cultivable bacteria from root canals in vivo and compared it with that of standard syringe irrigation (control group). Data were obtained by using bacteriologic sampling of root canals treated by endodontic residents. Sampling results from 1 session of treatment were then compared with results obtained after intervisit calcium hydroxide disinfection and a second session of treatment. There was no significant difference in the ability of sonic group and control group to eliminate cultivable bacteria from root canals (P > .05). A second session and intervisit calcium hydroxide disinfection were able to eliminate cultivable bacteria from significantly more teeth than a single session of treatment (P < .05). These in vivo results strengthen the case for a multi-visit approach to the treatment of apical periodontitis. Copyright 2010 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.
Sakai, Miho; Hayakawa, Yoshihiro; Funada, Yasuhiro; Ando, Takashi; Fukusaki, Eiichiro; Bamba, Takeshi
2017-09-15
In this study, we propose a novel variable sample injection system based on full-loop injection, named the split-flow sample introduction system, for application in supercritical fluid chromatography (SFC). In this system, the mobile phase is split by the differential pressure between two back pressure regulators (BPRs) after full-loop injection suitable for SFC, and this differential pressure determines the introduction rate. Nine compounds with a wide range of characteristics were introduced with high reproducibility and universality, confirming that a robust variable sample injection system was achieved. We also investigated the control factors of our proposed system. Sample introduction was controlled by the ratio between the column-side pressure drops in splitless and split flow, ΔP columnsideinsplitless and ΔP columnsideinsplit , respectively, where ΔP columnsideinsplitless is related to the mobile phase flow rate and composition and the column resistance. When all other conditions are kept constant, increasing the make-up flow induces an additional pressure drop on the column side of the system, which leads to a reduced column-side flow rate, and hence decreased the amount of sample injected, even when the net pressure drop on the column side remains the same. Thus, sample introduction could be highly controlled at low sample introduction rate, regardless of the introduction conditions. This feature is advantageous because, as a control factor, the solvent in the make-up pump is independent of the column-side pressure drop. Copyright © 2017. Published by Elsevier B.V.
Bertrand-Krajewski, J L; Bardin, J P; Mourad, M; Béranger, Y
2003-01-01
Assessing the functioning and the performance of urban drainage systems on both rainfall event and yearly time scales is usually based on online measurements of flow rates and on samples of influent effluent for some rainfall events per year. In order to draw pertinent scientific and operational conclusions from the measurement results, it is absolutely necessary to use appropriate methods and techniques in order to i) calibrate sensors and analytical methods, ii) validate raw data, iii) evaluate measurement uncertainties, iv) evaluate the number of rainfall events to sample per year in order to determine performance indicator with a given uncertainty. Based an previous work, the paper gives a synthetic review of required and techniques, and illustrates their application to storage and settling tanks. Experiments show that, controlled and careful experimental conditions, relative uncertainties are about 20% for flow rates in sewer pipes, 6-10% for volumes, 25-35% for TSS concentrations and loads, and 18-276% for TSS removal rates. In order to evaluate the annual pollutant interception efficiency of storage and settling tanks with a given uncertainty, efforts should first be devoted to decrease the sampling uncertainty by increasing the number of sampled events.
Lico, M.S.; Welch, A.H.; Hughes, J.L.
1986-01-01
The U.S. Geological Survey collected an extensive amount of hydrogeologic data from the shallow alluvial aquifer at two study sites near Fallon, Nevada, from 1984 though 1985. These data were collected as part of a study to determine the geochemical controls on the mobility of arsenic and other trace elements in shallow groundwater systems. The main study area is approximately 7 miles south of Fallon. A subsidiary study area is about 8 miles east of Fallon. The data collected include lithologic logs and water level altitudes for the augered sampling wells and piezometers, and determinations of arsenic and selenium content, grain size, porosity, hydraulic conductivity, and mineralogy for sediment samples from cores. (USGS)
NASA Astrophysics Data System (ADS)
Lora-Millán, Julio S.; Taguas, Encarnacion V.; Gomez, Jose A.; Perez, Rafael
2014-05-01
Long term soil erosion studies imply substantial efforts, particularly when there is the need to maintain continuous measurements. There are high costs associated to maintenance of field equipment keeping and quality control of data collection. Energy supply and/or electronic failures, vandalism and burglary are common causes of gaps in datasets, reducing their reach in many cases. In this work, a system of three video-cameras, a recorder and a transmission modem (3G technology) has been set up in a gauging station where rainfall, runoff flow and sediment concentration are monitored. The gauging station is located in the outlet of an olive orchard catchment of 6.4 ha. Rainfall is measured with one automatic raingauge that records intensity at one minute intervals. The discharge is measured by a flume of critical flow depth, where the water is recorded by an ultrasonic sensor. When the water level rises to a predetermined level, the automatic sampler turns on and fills a bottle at different intervals according to a program depending on the antecedent precipitation. A data logger controls the instruments' functions and records the data. The purpose of the video-camera system is to improve the quality of the dataset by i) the visual analysis of the measurement conditions of flow into the flume; ii) the optimisation of the sampling programs. The cameras are positioned to record the flow at the approximation and the gorge of the flume. In order to contrast the values of ultrasonic sensor, there is a third camera recording the flow level close to a measure tape. This system is activated when the ultrasonic sensor detects a height threshold, equivalent to an electric intensity level. Thus, only when there is enough flow, video-cameras record the event. This simplifies post-processing and reduces the cost of download of recordings. The preliminary contrast analysis will be presented as well as the main improvements in the sample program.
NASA Astrophysics Data System (ADS)
Wright, S. A.; Bennett, G. E.; Andrews, T.; Melis, T. S.; Topping, D. J.
2005-05-01
Currently, in-situ monitoring of water quality parameters (e.g. water temperature, conductivity, turbidity) in the Colorado River ecosystem typically consists of deploying instruments in the river, retrieving them at a later date, downloading the datalogger, then examining the data; an arduous process in the remote settings of Grand Canyon. Under this protocol, data is not available real-time and there is no way to detect problems with the instrumentation until after retrieval. The next obvious stage in the development of in-situ monitoring in Grand Canyon was the advent of one-way telemetry, i.e. streaming data in real-time from the instrument to the office and/or the world-wide-web. This protocol allows for real-time access to data and the identification of instrumentation problems, but still requires a site visit to address instrument malfunctions, i.e. the user does not have the ability to remotely control the instrument. At some field sites, such as the Colorado River in Grand Canyon, site visitation is restricted by remoteness and lack of traditional access routes (i.e. roads). Even at less remote sites, it may still be desirable to have two-way communication with instruments in order to, for example, diagnose and potentially fix instrumentation problems, change sampling parameters to save battery power, etc., without having to visit the site. To this end, the U.S. Geological Survey, Grand Canyon Monitoring and Research Center, is currently developing and testing a high-speed, two-way communication system that allows for real-time data access and remote control of instrumentation. The approach tested relies on internet access and may be especially useful in areas where land-line or cellular connections are unavailable. The system is composed of off-the-shelf products, uses a commercial broadband satellite service, and is designed in a generic way such that any instrument that communicates through RS-232 communication (i.e. a serial port) is compatible with the system. We are currently testing the system at two sites on the Colorado River in Grand Canyon and at one critical monitoring site on the Paria River where we have deployed suites of instruments for monitoring flow, sediment concentration, temperature, and conductivity. One aspect of the system that may be particularly useful for ecohydrological applications is the ability to remotely control on-site pump samplers, which allows for the collection of a water sample by the press of a button in the office.
Remote Monitoring of the Structural Health of Hydrokinetic Composite Turbine Blades
DOE Office of Scientific and Technical Information (OSTI.GOV)
J.L. Rovey
A health monitoring approach is investigated for hydrokinetic turbine blade applications. In-service monitoring is critical due to the difficult environment for blade inspection and the cost of inspection downtime. Composite blade designs have advantages that include long life in marine environments and great control over mechanical properties. Experimental strain characteristics are determined for static loads and free-vibration loads. These experiments are designed to simulate the dynamic characteristics of hydrokinetic turbine blades. Carbon/epoxy symmetric composite laminates are manufactured using an autoclave process. Four-layer composite beams, eight-layer composite beams, and two-dimensional eight-layer composite blades are instrumented for strain. Experimental results for strainmore » measurements from electrical resistance gages are validated with theoretical characteristics obtained from in-house finite-element analysis for all sample cases. These preliminary tests on the composite samples show good correlation between experimental and finite-element strain results. A health monitoring system is proposed in which damage to a composite structure, e.g. delamination and fiber breakage, causes changes in the strain signature behavior. The system is based on embedded strain sensors and embedded motes in which strain information is demodulated for wireless transmission. In-service monitoring is critical due to the difficult environment for blade inspection and the cost of inspection downtime. Composite blade designs provide a medium for embedding sensors into the blades for in-situ health monitoring. The major challenge with in-situ health monitoring is transmission of sensor signals from the remote rotating reference frame of the blade to the system monitoring station. In the presented work, a novel system for relaying in-situ blade health measurements in hydrokinetic systems is described and demonstrated. An ultrasonic communication system is used to transmit sensor data underwater from the rotating frame of the blade to a fixed relay station. Data are then broadcast via radio waves to a remote monitoring station. Results indicate that the assembled system can transmit simulated sensor data with an accuracy of ±5% at a maximum sampling rate of 500 samples/sec. A power investigation of the transmitter within the blade shows that continuous max-sampling operation is only possible for short durations (~days), and is limited due to the capacity of the battery power source. However, intermittent sampling, with long periods between samples, allows for the system to last for very long durations (~years). Finally, because the data transmission system can operate at a high sampling rate for short durations or at a lower sampling rate/higher duty cycle for long durations, it is well-suited for short-term prototype and environmental testing, as well as long-term commercially-deployed hydrokinetic machines.« less
NASA Astrophysics Data System (ADS)
Chan, S.; Lehnert, K. A.; Coleman, R. J.
2011-12-01
SESAR, the System for Earth Sample Registration, is an online registry for physical samples collected for Earth and environmental studies. SESAR generates and administers the International Geo Sample Number IGSN, a unique identifier for samples that is dramatically advancing interoperability amongst information systems for sample-based data. SESAR was developed to provide the complete range of registry services, including definition of IGSN syntax and metadata profiles, registration and validation of name spaces requested by users, tools for users to submit and manage sample metadata, validation of submitted metadata, generation and validation of the unique identifiers, archiving of sample metadata, and public or private access to the sample metadata catalog. With the development of SESAR v3, we placed particular emphasis on creating enhanced tools that make metadata submission easier and more efficient for users, and that provide superior functionality for users to manage metadata of their samples in their private workspace MySESAR. For example, SESAR v3 includes a module where users can generate custom spreadsheet templates to enter metadata for their samples, then upload these templates online for sample registration. Once the content of the template is uploaded, it is displayed online in an editable grid format. Validation rules are executed in real-time on the grid data to ensure data integrity. Other new features of SESAR v3 include the capability to transfer ownership of samples to other SESAR users, the ability to upload and store images and other files in a sample metadata profile, and the tracking of changes to sample metadata profiles. In the next version of SESAR (v3.5), we will further improve the discovery, sharing, registration of samples. For example, we are developing a more comprehensive suite of web services that will allow discovery and registration access to SESAR from external systems. Both batch and individual registrations will be possible through web services. Based on valuable feedback from the user community, we will introduce enhancements that add greater flexibility to the system to accommodate the vast diversity of metadata that users want to store. Users will be able to create custom metadata fields and use these for the samples they register. Users will also be able to group samples into 'collections' to make retrieval for research projects or publications easier. An improved interface design will allow for better workflow transition and navigation throughout the application. In keeping up with the demands of a growing community, SESAR has also made process changes to ensure efficiency in system development. For example, we have implemented a release cycle to better track enhancements and fixes to the system, and an API library that facilitates reusability of code. Usage tracking, metrics and surveys capture information to guide the direction of future developments. A new set of administrative tools allows greater control of system management.
A novel pulsed gas metal arc welding system with direct droplet transfer close-loop control
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang, Q.; Li, P.; Zhang, L.
1994-12-31
In pulsed gas metal arc welding (GMAW), a predominant parameter that has to be monitored and controlled in real time for maintaining process stability and ensuring weld quality, is droplet transfer. Based on the close correlation between droplet transfer and arc light radiant flux in GMAW of steel and aluminum, a direct closed-loop droplet transfer control system for pulsed GMAW with arc light sensor has been developed. By sensing the droplet transfer directly via the arc light signal, a pulsed GMAW process with real and exact one-pulse, one-droplet transfer has been achieved. The novel pulsed GMAW machine consists of threemore » parts: a sensing system, a controlling system, and a welding power system. The software used in this control system is capable of data sampling and processing, parameter matching, optimum parameter restoring, and resetting. A novel arc light sensing system has been developed. The sensor is small enough to be clamped to a semiautomatic welding torch. Based on thissensingn system, a closed-loop droplet transfer control system of GMAW of steel and aluminum has been built and a commercial prototype has been made. The system is capable of keeping one-pulse, one-droplet transfer against external interferences. The welding process with this control system has been proved to be stable, quiet, with no spatter, and provide good weld formation.« less
An Overview of Recent Advances in Event-Triggered Consensus of Multiagent Systems.
Ding, Lei; Han, Qing-Long; Ge, Xiaohua; Zhang, Xian-Ming
2018-04-01
Event-triggered consensus of multiagent systems (MASs) has attracted tremendous attention from both theoretical and practical perspectives due to the fact that it enables all agents eventually to reach an agreement upon a common quantity of interest while significantly alleviating utilization of communication and computation resources. This paper aims to provide an overview of recent advances in event-triggered consensus of MASs. First, a basic framework of multiagent event-triggered operational mechanisms is established. Second, representative results and methodologies reported in the literature are reviewed and some in-depth analysis is made on several event-triggered schemes, including event-based sampling schemes, model-based event-triggered schemes, sampled-data-based event-triggered schemes, and self-triggered sampling schemes. Third, two examples are outlined to show applicability of event-triggered consensus in power sharing of microgrids and formation control of multirobot systems, respectively. Finally, some challenging issues on event-triggered consensus are proposed for future research.
Peterson, Alexis B; Sauber-Schatz, Erin K; Mack, Karin A
2018-06-01
As more states legalize medical/recreational marijuana use, it is important to determine if state motor-vehicle surveillance systems can effectively monitor and track driving under the influence (DUI) of marijuana. This study assessed Colorado's Department of Revenue motor-vehicle crash data system, Electronic Accident Reporting System (EARS), to monitor non-fatal crashes involving driving under the influence (DUI) of marijuana. Centers for Disease Control and Prevention guidelines on surveillance system evaluation were used to assess EARS' usefulness, flexibility, timeliness, simplicity, acceptability, and data quality. We assessed system components, interviewed key stakeholders, and analyzed completeness of Colorado statewide 2014 motor-vehicle crash records. EARS contains timely and complete data, but does not effectively monitor non-fatal motor-vehicle crashes related to DUI of marijuana. Information on biological sample type collected from drivers and toxicology results were not recorded into EARS; however, EARS is a flexible system that can incorporate new data without increasing surveillance system burden. States, including Colorado, could consider standardization of drug testing and mandatory reporting policies for drivers involved in motor-vehicle crashes and proactively address the narrow window of time for sample collection to improve DUI of marijuana surveillance. Practical applications: The evaluation of state motor-vehicle crash systems' ability to capture crashes involving drug impaired driving (DUID) is a critical first step for identifying frequency and risk factors for crashes related to DUID. Published by Elsevier Ltd.
40 CFR 63.1013 - Sampling connection systems standards.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 40 Protection of Environment 10 2010-07-01 2010-07-01 false Sampling connection systems standards... (CONTINUED) National Emission Standards for Equipment Leaks-Control Level 1 § 63.1013 Sampling connection... sampling connection system shall be equipped with a closed purge, closed loop, or closed vent system...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Castaldini, C.; Waterland, L.R.
1987-03-01
The two-volume report gives results from field tests of a wood-waste-fired industrial watertube boiler. Two series of tests were performed: one firing dry (11% moisture) wood waste, and the other firing green (34% moisture) wood waste. Emission measurements included: continuous monitoring of flue-gas emissions; source-assessment sampling system (SASS) sampling of the flue-gas with subsequent laboratory analysis of samples to give total flue-gas organics in two boiling-point ranges, compound category information within these ranges, specific quantitation of the semi-volatile organic priority pollutants, and flue gas concentrations of 73 trace elements; Method 5 sampling for particulate; controlled condensation system sampling for SO/submore » 2/ and SO/sub 3/; and grab sampling of boiler mechanical collector hopper ash for inorganic and organic composition determinations. Total organic emissions decreased from 60-135 mg/dscm firing dry wood to 2-65 mg/dscm firing green wood, in parallel with corresponding boiler CO emissions.« less
Nano-metal oxides: Exposure and engineering control assessment.
Garcia, Alberto; Eastlake, Adrienne; Topmiller, Jennifer L; Sparks, Christopher; Martinez, Kenneth; Geraci, Charles L
2017-09-01
In January 2007, the National Institute for Occupational Safety and Health (NIOSH) conducted a field study to evaluate process specific emissions during the production of ENMs. This study was performed using the nanoparticle emission assessment technique (NEAT). During this study, it was determined that ENMs were released during production and cleaning of the process reactor. Airborne concentrations of silver, nickel, and iron were found both in the employee's personal breathing zone and area samples during reactor cleaning. At the completion of this initial survey, it was suggested that a flanged attachment be added to the local exhaust ventilation system. NIOSH re-evaluated the facility in December 2011 to assess worker exposures following an increase in production rates. This study included a fully comprehensive emissions, exposure, and engineering control evaluation of the entire process. This study made use of the nanoparticle exposure assessment technique (NEAT 2.0). Data obtained from filter-based samples and direct reading instruments indicate that reactor cleanout increased the overall particle concentration in the immediate area. However, it does not appear that these concentrations affect areas outside of the production floor. As the distance between the reactor and the sample location increased, the observed particle number concentration decreased, creating a concentration gradient with respect to the reactor. The results of this study confirm that the flanged attachment on the local exhaust ventilation system served to decrease exposure potential. Given the available toxicological data of the metals evaluated, caution is warranted. One should always keep in mind that occupational exposure levels were not developed specifically for nanoscale particles. With data suggesting that certain nanoparticles may be more toxic than the larger counterparts of the same material; employers should attempt to control emissions of these particles at the source, to limit the potential for exposure.
de Beer, R; Graveron-Demilly, D; Nastase, S; van Ormondt, D
2004-03-01
Recently we have developed a Java-based heterogeneous distributed computing system for the field of magnetic resonance imaging (MRI). It is a software system for embedding the various image reconstruction algorithms that we have created for handling MRI data sets with sparse sampling distributions. Since these data sets may result from multi-dimensional MRI measurements our system has to control the storage and manipulation of large amounts of data. In this paper we describe how we have employed the extensible markup language (XML) to realize this data handling in a highly structured way. To that end we have used Java packages, recently released by Sun Microsystems, to process XML documents and to compile pieces of XML code into Java classes. We have effectuated a flexible storage and manipulation approach for all kinds of data within the MRI system, such as data describing and containing multi-dimensional MRI measurements, data configuring image reconstruction methods and data representing and visualizing the various services of the system. We have found that the object-oriented approach, possible with the Java programming environment, combined with the XML technology is a convenient way of describing and handling various data streams in heterogeneous distributed computing systems.
NASA Astrophysics Data System (ADS)
Meng, Su; Chen, Jie; Sun, Jian
2017-10-01
This paper investigates the problem of observer-based output feedback control for networked control systems with non-uniform sampling and time-varying transmission delay. The sampling intervals are assumed to vary within a given interval. The transmission delay belongs to a known interval. A discrete-time model is first established, which contains time-varying delay and norm-bounded uncertainties coming from non-uniform sampling intervals. It is then converted to an interconnection of two subsystems in which the forward channel is delay-free. The scaled small gain theorem is used to derive the stability condition for the closed-loop system. Moreover, the observer-based output feedback controller design method is proposed by utilising a modified cone complementary linearisation algorithm. Finally, numerical examples illustrate the validity and superiority of the proposed method.
Micro injector sample delivery system for charged molecules
Davidson, James C.; Balch, Joseph W.
1999-11-09
A micro injector sample delivery system for charged molecules. The injector is used for collecting and delivering controlled amounts of charged molecule samples for subsequent analysis. The injector delivery system can be scaled to large numbers (>96) for sample delivery to massively parallel high throughput analysis systems. The essence of the injector system is an electric field controllable loading tip including a section of porous material. By applying the appropriate polarity bias potential to the injector tip, charged molecules will migrate into porous material, and by reversing the polarity bias potential the molecules are ejected or forced away from the tip. The invention has application for uptake of charged biological molecules (e.g. proteins, nucleic acids, polymers, etc.) for delivery to analytical systems, and can be used in automated sample delivery systems.
NASA Astrophysics Data System (ADS)
Arnold, F.; DeMallie, I.; Florence, L.; Kashinski, D. O.
2015-03-01
This manuscript addresses the design, hardware details, construction, and programming of an apparatus allowing an experimenter to monitor and record high-temperature thermocouple measurements of dynamic systems in real time. The apparatus uses wireless network technology to bridge the gap between a dynamic (moving) sample frame and the static laboratory frame. Our design is a custom solution applied to samples that rotate through large angular displacements where hard-wired and typical slip-ring solutions are not practical because of noise considerations. The apparatus consists of a Raspberry PI mini-Linux computer, an Arduino micro-controller, an Ocean Controls thermocouple multiplexer shield, and k-type thermocouples.
Arnold, F; DeMallie, I; Florence, L; Kashinski, D O
2015-03-01
This manuscript addresses the design, hardware details, construction, and programming of an apparatus allowing an experimenter to monitor and record high-temperature thermocouple measurements of dynamic systems in real time. The apparatus uses wireless network technology to bridge the gap between a dynamic (moving) sample frame and the static laboratory frame. Our design is a custom solution applied to samples that rotate through large angular displacements where hard-wired and typical slip-ring solutions are not practical because of noise considerations. The apparatus consists of a Raspberry PI mini-Linux computer, an Arduino micro-controller, an Ocean Controls thermocouple multiplexer shield, and k-type thermocouples.
Data-driven sensor placement from coherent fluid structures
NASA Astrophysics Data System (ADS)
Manohar, Krithika; Kaiser, Eurika; Brunton, Bingni W.; Kutz, J. Nathan; Brunton, Steven L.
2017-11-01
Optimal sensor placement is a central challenge in the prediction, estimation and control of fluid flows. We reinterpret sensor placement as optimizing discrete samples of coherent fluid structures for full state reconstruction. This permits a drastic reduction in the number of sensors required for faithful reconstruction, since complex fluid interactions can often be described by a small number of coherent structures. Our work optimizes point sensors using the pivoted matrix QR factorization to sample coherent structures directly computed from flow data. We apply this sampling technique in conjunction with various data-driven modal identification methods, including the proper orthogonal decomposition (POD) and dynamic mode decomposition (DMD). In contrast to POD-based sensors, DMD demonstrably enables the optimization of sensors for prediction in systems exhibiting multiple scales of dynamics. Finally, reconstruction accuracy from pivot sensors is shown to be competitive with sensors obtained using traditional computationally prohibitive optimization methods.
Aquatic plant control research
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pryfogle, P.A.; Rinehart, B.N.; Ghio, E.G.
The Northwest region of the United States contains extensive canal systems that transport water for hydropower generation. Nuisance plants, including algae, that grow in these systems reduce their hydraulic capacity through water displacement and increased surface friction. Most control methods are applied in an ad hoc fashion. The goal of this work is to develop cost-effective, environmentally sound, long-term management strategies to prevent and control nuisance algal growth. This paper reports on a multi-year study, performed in collaboration with the Pacific Gas & Electric Company, to investigate algal growth in their canal systems, and to evaluate various control methodologies. Threemore » types of controls, including mechanical, biological and chemical treatment, were selected for testing and evaluation. As part of this study, water quality data were collected and algal communities were sampled from numerous stations throughout the distribution system at regular intervals. This study resulted in a more comprehensive understanding of conditions leading to the development of nuisance algal growth, a better informed selection of treatment plans, and improved evaluation of the effectiveness for the control strategies selected for testing.« less
Multichannel infrared fiber optic radiometer for controlled microwave heating
NASA Astrophysics Data System (ADS)
Drizlikh, S.; Zur, Albert; Katzir, Abraham
1990-07-01
An infrared fiberoptic multichannel radiometer was used for monitoring and controlling the temperature of samples in a microwave heating system. The temperature of water samples was maintained at about 40 °C, with a standard deviation of +/- 0.2°C and a maximum deviation of +/- 0.5°C. The temperature was monitored on the same time at several points on the surface and inside the sample. This novel controlled system is reliable and precise. Such system would be very useful for medical applications such as hypothermia and hyperthermi a.
DOE Office of Scientific and Technical Information (OSTI.GOV)
A. Alfonsi; C. Rabiti; D. Mandelli
The Reactor Analysis and Virtual control ENviroment (RAVEN) code is a software tool that acts as the control logic driver and post-processing engine for the newly developed Thermal-Hydraulic code RELAP-7. RAVEN is now a multi-purpose Probabilistic Risk Assessment (PRA) software framework that allows dispatching different functionalities: Derive and actuate the control logic required to simulate the plant control system and operator actions (guided procedures), allowing on-line monitoring/controlling in the Phase Space Perform both Monte-Carlo sampling of random distributed events and Dynamic Event Tree based analysis Facilitate the input/output handling through a Graphical User Interface (GUI) and a post-processing data miningmore » module« less
Use of multivariate statistics to identify unreliable data obtained using CASA.
Martínez, Luis Becerril; Crispín, Rubén Huerta; Mendoza, Maximino Méndez; Gallegos, Oswaldo Hernández; Martínez, Andrés Aragón
2013-06-01
In order to identify unreliable data in a dataset of motility parameters obtained from a pilot study acquired by a veterinarian with experience in boar semen handling, but without experience in the operation of a computer assisted sperm analysis (CASA) system, a multivariate graphical and statistical analysis was performed. Sixteen boar semen samples were aliquoted then incubated with varying concentrations of progesterone from 0 to 3.33 µg/ml and analyzed in a CASA system. After standardization of the data, Chernoff faces were pictured for each measurement, and a principal component analysis (PCA) was used to reduce the dimensionality and pre-process the data before hierarchical clustering. The first twelve individual measurements showed abnormal features when Chernoff faces were drawn. PCA revealed that principal components 1 and 2 explained 63.08% of the variance in the dataset. Values of principal components for each individual measurement of semen samples were mapped to identify differences among treatment or among boars. Twelve individual measurements presented low values of principal component 1. Confidence ellipses on the map of principal components showed no statistically significant effects for treatment or boar. Hierarchical clustering realized on two first principal components produced three clusters. Cluster 1 contained evaluations of the two first samples in each treatment, each one of a different boar. With the exception of one individual measurement, all other measurements in cluster 1 were the same as observed in abnormal Chernoff faces. Unreliable data in cluster 1 are probably related to the operator inexperience with a CASA system. These findings could be used to objectively evaluate the skill level of an operator of a CASA system. This may be particularly useful in the quality control of semen analysis using CASA systems.
Wagenaar, Bradley H; Sherr, Kenneth; Fernandes, Quinhas; Wagenaar, Alexander C
2016-01-01
Routine health information systems (RHISs) are in place in nearly every country and provide routinely collected full-coverage records on all levels of health system service delivery. However, these rich sources of data are regularly overlooked for evaluating causal effects of health programmes due to concerns regarding completeness, timeliness, representativeness and accuracy. Using Mozambique’s national RHIS (Módulo Básico) as an illustrative example, we urge renewed attention to the use of RHIS data for health evaluations. Interventions to improve data quality exist and have been tested in low-and middle-income countries (LMICs). Intrinsic features of RHIS data (numerous repeated observations over extended periods of time, full coverage of health facilities, and numerous real-time indicators of service coverage and utilization) provide for very robust quasi-experimental designs, such as controlled interrupted time-series (cITS), which are not possible with intermittent community sample surveys. In addition, cITS analyses are well suited for continuously evolving development contexts in LMICs by: (1) allowing for measurement and controlling for trends and other patterns before, during and after intervention implementation; (2) facilitating the use of numerous simultaneous control groups and non-equivalent dependent variables at multiple nested levels to increase validity and strength of causal inference; and (3) allowing the integration of continuous ‘effective dose received’ implementation measures. With expanded use of RHIS data for the evaluation of health programmes, investments in data systems, health worker interest in and utilization of RHIS data, as well as data quality will further increase over time. Because RHIS data are ministry-owned and operated, relying upon these data will contribute to sustainable national capacity over time. PMID:25887561
Wagenaar, Bradley H; Sherr, Kenneth; Fernandes, Quinhas; Wagenaar, Alexander C
2016-02-01
Routine health information systems (RHISs) are in place in nearly every country and provide routinely collected full-coverage records on all levels of health system service delivery. However, these rich sources of data are regularly overlooked for evaluating causal effects of health programmes due to concerns regarding completeness, timeliness, representativeness and accuracy. Using Mozambique's national RHIS (Módulo Básico) as an illustrative example, we urge renewed attention to the use of RHIS data for health evaluations. Interventions to improve data quality exist and have been tested in low-and middle-income countries (LMICs). Intrinsic features of RHIS data (numerous repeated observations over extended periods of time, full coverage of health facilities, and numerous real-time indicators of service coverage and utilization) provide for very robust quasi-experimental designs, such as controlled interrupted time-series (cITS), which are not possible with intermittent community sample surveys. In addition, cITS analyses are well suited for continuously evolving development contexts in LMICs by: (1) allowing for measurement and controlling for trends and other patterns before, during and after intervention implementation; (2) facilitating the use of numerous simultaneous control groups and non-equivalent dependent variables at multiple nested levels to increase validity and strength of causal inference; and (3) allowing the integration of continuous 'effective dose received' implementation measures. With expanded use of RHIS data for the evaluation of health programmes, investments in data systems, health worker interest in and utilization of RHIS data, as well as data quality will further increase over time. Because RHIS data are ministry-owned and operated, relying upon these data will contribute to sustainable national capacity over time. © The Author 2015. Published by Oxford University Press in association with The London School of Hygiene and Tropical Medicine.
NASA Technical Reports Server (NTRS)
Halyo, N.
1983-01-01
The design and development of a 3-D Digital Integrated Automatic Landing System (DIALS) for the Terminal Configured Vehicle (TCV) Research Aircraft, a B-737-100 is described. The system was designed using sampled data Linear Quadratic Gaussian (LOG) methods, resulting in a direct digital design with a modern control structure which consists of a Kalman filter followed by a control gain matrix, all operating at 10 Hz. DIALS uses Microwave Landing System (MLS) position, body-mounted accelerometers, as well as on-board sensors usually available on commercial aircraft, but does not use inertial platforms. The phases of the final approach considered are the localizer and glideslope capture which may be performed simultaneously, localizer and steep glideslope track or hold, crab/decrab and flare to touchdown. DIALS captures, tracks and flares from steep glideslopes ranging from 2.5 deg to 5.5 deg, selected prior to glideslope capture. Digital Integrated Automatic Landing System is the first modern control design automatic landing system successfully flight tested. The results of an initial nonlinear simulation are presented here.
Automation of Command and Data Entry in a Glovebox Work Volume: An Evaluation of Data Entry Devices
NASA Technical Reports Server (NTRS)
Steele, Marianne K.; Nakamura, Gail; Havens, Cindy; LeMay, Moira
1996-01-01
The present study was designed to examine the human-computer interface for data entry while performing experimental procedures within a glovebox work volume in order to make a recommendation to the Space Station Biological Research Project for a data entry system to be used within the Life Sciences Glovebox. Test subjects entered data using either a manual keypad, similar to a standard computer numerical keypad located within the glovebox work volume, or a voice input system using a speech recognition program with a microphone headset. Numerical input and commands were programmed in an identical manner between the two systems. With both electronic systems, a small trackball was available within the work volume for cursor control. Data, such as sample vial identification numbers, sample tissue weights, and health check parameters of the specimen, were entered directly into procedures that were electronically displayed on a video monitor within the glovebox. A pen and paper system with a 'flip-chart' format for procedure display, similar to that currently in use on the Space Shuttle, was used as a baseline data entry condition. Procedures were performed by a single operator; eight test subjects were used in the study. The electronic systems were tested under both a 'nominal' or 'anomalous' condition. The anomalous condition was introduced into the experimental procedure to increase the probability of finding limitations or problems with human interactions with the electronic systems. Each subject performed five test runs during a test day: two procedures each with voice and keypad, one with and one without anomalies, and one pen and paper procedure. The data collected were both quantitative (times, errors) and qualitative (subjective ratings of the subjects).
Zhang, Cuicui; Liang, Xuefeng; Matsuyama, Takashi
2014-12-08
Multi-camera networks have gained great interest in video-based surveillance systems for security monitoring, access control, etc. Person re-identification is an essential and challenging task in multi-camera networks, which aims to determine if a given individual has already appeared over the camera network. Individual recognition often uses faces as a trial and requires a large number of samples during the training phrase. This is difficult to fulfill due to the limitation of the camera hardware system and the unconstrained image capturing conditions. Conventional face recognition algorithms often encounter the "small sample size" (SSS) problem arising from the small number of training samples compared to the high dimensionality of the sample space. To overcome this problem, interest in the combination of multiple base classifiers has sparked research efforts in ensemble methods. However, existing ensemble methods still open two questions: (1) how to define diverse base classifiers from the small data; (2) how to avoid the diversity/accuracy dilemma occurring during ensemble. To address these problems, this paper proposes a novel generic learning-based ensemble framework, which augments the small data by generating new samples based on a generic distribution and introduces a tailored 0-1 knapsack algorithm to alleviate the diversity/accuracy dilemma. More diverse base classifiers can be generated from the expanded face space, and more appropriate base classifiers are selected for ensemble. Extensive experimental results on four benchmarks demonstrate the higher ability of our system to cope with the SSS problem compared to the state-of-the-art system.
Zhang, Cuicui; Liang, Xuefeng; Matsuyama, Takashi
2014-01-01
Multi-camera networks have gained great interest in video-based surveillance systems for security monitoring, access control, etc. Person re-identification is an essential and challenging task in multi-camera networks, which aims to determine if a given individual has already appeared over the camera network. Individual recognition often uses faces as a trial and requires a large number of samples during the training phrase. This is difficult to fulfill due to the limitation of the camera hardware system and the unconstrained image capturing conditions. Conventional face recognition algorithms often encounter the “small sample size” (SSS) problem arising from the small number of training samples compared to the high dimensionality of the sample space. To overcome this problem, interest in the combination of multiple base classifiers has sparked research efforts in ensemble methods. However, existing ensemble methods still open two questions: (1) how to define diverse base classifiers from the small data; (2) how to avoid the diversity/accuracy dilemma occurring during ensemble. To address these problems, this paper proposes a novel generic learning-based ensemble framework, which augments the small data by generating new samples based on a generic distribution and introduces a tailored 0–1 knapsack algorithm to alleviate the diversity/accuracy dilemma. More diverse base classifiers can be generated from the expanded face space, and more appropriate base classifiers are selected for ensemble. Extensive experimental results on four benchmarks demonstrate the higher ability of our system to cope with the SSS problem compared to the state-of-the-art system. PMID:25494350
NASA Astrophysics Data System (ADS)
Fischer, Philipp; Schwanitz, Max; Loth, Reiner; Posner, Uwe; Brand, Markus; Schröder, Friedhelm
2017-04-01
A combined year-round assessment of selected oceanographic data and a macrobiotic community assessment was performed from October 2013 to November 2014 in the littoral zone of the Kongsfjorden polar fjord system on the western coast of Svalbard (Norway). State of the art remote controlled cabled underwater observatory technology was used for daily vertical profiles of temperature, salinity, and turbidity together with a stereo-optical assessment of the macrobiotic community, including fish. The results reveal a distinct seasonal cycle in total species abundances, with a significantly higher total abundance and species richness during the polar winter when no light is available underwater compared to the summer months when 24 h light is available. During the winter months, a temporally highly segmented community was observed with respect to species occurrence, with single species dominating the winter community for restricted times. In contrast, the summer community showed an overall lower total abundance as well as a significantly lower number of species. The study clearly demonstrates the high potential of cable connected remote controlled digital sampling devices, especially in remote areas, such as polar fjord systems, with harsh environmental conditions and limited accessibility. A smart combination of such new digital sampling
methods with classic sampling procedures can provide a possibility to significantly extend the sampling time and frequency, especially in remote and difficult to access areas. This can help to provide a sufficient data density and therefore statistical power for a sound scientific analysis without increasing the invasive sampling pressure in ecologically sensitive environments.
NASA Astrophysics Data System (ADS)
Wardi, Jeni; Yandra, Alexsander
2018-05-01
This research aims to learn the direct influence of transformational and transactional leaderships on Indonesian company’s performance through company’s culture, strategy, management accounting and control system as a marine state. This research involves descriptive and inferential designs in solving the research problem. To test the model and the hypothesis, SEM analysis is used. The populations of this research are companies registered in Indonesian stock exchange in 2012. The sampling technique uses purposive sampling. The data of the research are obtained from questionnaires distributed to respondents. The respondents are companies’ managers represented by accounting and finance managers with the positions 1 and 2 levels below top management team who have direct communication with the top management. The results of the research show that transformational leadership influences company’s performance directly, but not the transactional leadership. The company’s culture is not the mediation variable in indirect influence on the company’s performance, either in transformational or transactional leadership. On the other hand, management control system proves to be the mediation in transactional leadership on the performance but not for transformational leadership. Meanwhile, management accounting system proves to be the mediation variable in the influence of transformational and transactional leaderships. Except the variables of company’s culture, strategy, management accounting system and management control system, each directly influences the performance.
Interactive water monitoring system accessible by cordless telephone
NASA Astrophysics Data System (ADS)
Volpicelli, Richard; Andeweg, Pierre; Hagar, William G.
1985-12-01
A battery-operated, microcomputer-controlled monitoring device linked with a cordless telephone has been developed for remote measurements. This environmental sensor is self-contained and collects and processes data according to the information sent to its on-board computer system. An RCA model 1805 microprocessor forms the basic controller with a program encoded in memory for data acquisition and analysis. Signals from analog sensing devices used to monitor the environment are converted into digital signals and stored in random access memory of the microcomputer. This remote sensing system is linked to the laboratory by means of a cordless telephone whose base unit is connected to regular telephone lines. This offshore sensing system is simply accessed by a phone call originating from a computer terminal in the laboratory. Data acquisition is initiated upon request: Information continues to be processed and stored until the computer is reprogrammed by another phone call request. Information obtained may be recalled by a phone call after the desired environmental measurements are finished or while they are in progress. Data sampling parameters may be reset at any time, including in the middle of a measurement cycle. The range of the system is limited only by existing telephone grid systems and by the transmission characteristics of the cordless phone used as a communications link. This use of a cordless telephone, coupled with the on-board computer system, may be applied to other field studies requiring data transfer between an on-site analytical system and the laboratory.
NASA Technical Reports Server (NTRS)
Dickey, Tommy D.; Granata, Timothy C.; Taupier-Letage, Isabelle
1992-01-01
The processes controlling the flux of carbon in the upper ocean have dynamic ranges in space and time of at least nine orders of magnitude. These processes depend on a broad suite of inter-related biogeochemical, bio-optical, and physical variables. These variables should be sampled on scales matching the relevant phenomena. Traditional ship-based sampling, while critical for detailed and more comprehensive observations, can span only limited portions of these ranges because of logistical and financial constraints. Further, remote observations from satellite platforms enable broad horizontal coverage which is restricted to the upper few meters of the ocean. For these main reasons, automated subsurface measurement systems are important for the fulfillment of research goals related to the regional and global estimation and modeling of time varying biogeochemical fluxes. Within the past few years, new sensors and systems capable of autonomously measuring several of the critical variables have been developed. The platforms for deploying these systems now include moorings and drifters and it is likely that autonomous underwater vehicles (AUV's) will become available for use in the future. Each of these platforms satisfies particular sampling needs and can be used to complement both shipboard and satellite observations. In the present review, (1) sampling considerations will be summarized, (2) examples of data obtained from some of the existing automated in situ sampling systems will be highlighted, (3) future sensors and systems will be discussed, (4) data management issues for present and future automated systems will be considered, and (5) the status of near real-time data telemetry will be outlined. Finally, we wish to make it clear at the outset that the perspectives presented here are those of the authors and are not intended to represent those of the United States JGOFS program, the International JGOFS program, NOAA's C&GC program, or other global ocean programs.
Charlestra, Lucner; Amirbahman, Aria; Courtemanch, David L.; Alvarez, David A.; Patterson, Howard
2012-01-01
The polar organic chemical integrative sampler (POCIS) was calibrated to monitor pesticides in water under controlled laboratory conditions. The effect of natural organic matter (NOM) on the sampling rates (Rs) was evaluated in microcosms containing -1 of total organic carbon (TOC). The effect of hydrodynamics was studied by comparing Rs values measured in stirred (SBE) and quiescent (QBE) batch experiments and a flow-through system (FTS). The level of NOM in the water used in these experiments had no effect on the magnitude of the pesticide sampling rates (p > 0.05). However, flow velocity and turbulence significantly increased the sampling rates of the pesticides in the FTS and SBE compared to the QBE (p < 0.001). The calibration data generated can be used to derive pesticide concentrations in water from POCIS deployed in stagnant and turbulent environmental systems without correction for NOM.
A Control Chart Approach for Representing and Mining Data Streams with Shape Based Similarity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Omitaomu, Olufemi A
The mining of data streams for online condition monitoring is a challenging task in several domains including (electric) power grid system, intelligent manufacturing, and consumer science. Considering a power grid application in which thousands of sensors, called the phasor measurement units, are deployed on the power grid network to continuously collect streams of digital data for real-time situational awareness and system management. Depending on design, each sensor could stream between ten and sixty data samples per second. The myriad of sensory data captured could convey deeper insights about sequence of events in real-time and before major damages are done. However,more » the timely processing and analysis of these high-velocity and high-volume data streams is a challenge. Hence, a new data processing and transformation approach, based on the concept of control charts, for representing sequence of data streams from sensors is proposed. In addition, an application of the proposed approach for enhancing data mining tasks such as clustering using real-world power grid data streams is presented. The results indicate that the proposed approach is very efficient for data streams storage and manipulation.« less
SAMURAI: Polar AUV-Based Autonomous Dexterous Sampling
NASA Astrophysics Data System (ADS)
Akin, D. L.; Roberts, B. J.; Smith, W.; Roderick, S.; Reves-Sohn, R.; Singh, H.
2006-12-01
While autonomous undersea vehicles are increasingly being used for surveying and mapping missions, as of yet there has been little concerted effort to create a system capable of performing physical sampling or other manipulation of the local environment. This type of activity has typically been performed under teleoperated control from ROVs, which provides high-bandwidth real-time human direction of the manipulation activities. Manipulation from an AUV will require a completely autonomous sampling system, which implies both advanced technologies such as machine vision and autonomous target designation, but also dexterous robot manipulators to perform the actual sampling without human intervention. As part of the NASA Astrobiology Science and Technology for Exploring the Planets (ASTEP) program, the University of Maryland Space Systems Laboratory has been adapting and extending robotics technologies developed for spacecraft assembly and maintenance to the problem of autonomous sampling of biologicals and soil samples around hydrothermal vents. The Sub-polar ice Advanced Manipulator for Universal Sampling and Autonomous Intervention (SAMURAI) system is comprised of a 6000-meter capable six-degree-of-freedom dexterous manipulator, along with an autonomous vision system, multi-level control system, and sampling end effectors and storage mechanisms to allow collection of samples from vent fields. SAMURAI will be integrated onto the Woods Hole Oceanographic Institute (WHOI) Jaguar AUV, and used in Arctic during the fall of 2007 for autonomous vent field sampling on the Gakkel Ridge. Under the current operations concept, the JAGUAR and PUMA AUVs will survey the water column and localize on hydrothermal vents. Early mapping missions will create photomosaics of the vents and local surroundings, allowing scientists on the mission to designate desirable sampling targets. Based on physical characteristics such as size, shape, and coloration, the targets will be loaded into the SAMURAI control system, and JAGUAR (with SAMURAI mounted to the lower forward hull) will return to the designated target areas. Once on site, vehicle control will be turned over to the SAMURAI controller, which will perform vision-based guidance to the sampling site and will then ground the AUV to the sea bottom for stability. The SAMURAI manipulator will collect samples, such as sessile biologicals, geological samples, and (potentially) vent fluids, and store the samples for the return trip. After several hours of sampling operations on one or several sites, JAGUAR control will be returned to the WHOI onboard controller for the return to the support ship. (Operational details of AUV operations on the Gakkel Ridge mission are presented in other papers at this conference.) Between sorties, SAMURAI end effectors can be changed out on the surface for specific targets, such as push cores or larger biologicals such as tube worms. In addition to the obvious challenges in autonomous vision-based manipulator control from a free-flying support vehicle, significant development challenges have been the design of a highly capable robotic arm within the mass limitations (both wet and dry) of the JAGUAR vehicle, the development of a highly robust manipulator with modular maintenance units for extended polar operations, and the creation of a robot-based sample collection and holding system for multiple heterogeneous samples on a single extended sortie.
Noise and Complexity in Human Postural Control: Interpreting the Different Estimations of Entropy
Rhea, Christopher K.; Silver, Tobin A.; Hong, S. Lee; Ryu, Joong Hyun; Studenka, Breanna E.; Hughes, Charmayne M. L.; Haddad, Jeffrey M.
2011-01-01
Background Over the last two decades, various measures of entropy have been used to examine the complexity of human postural control. In general, entropy measures provide information regarding the health, stability and adaptability of the postural system that is not captured when using more traditional analytical techniques. The purpose of this study was to examine how noise, sampling frequency and time series length influence various measures of entropy when applied to human center of pressure (CoP) data, as well as in synthetic signals with known properties. Such a comparison is necessary to interpret data between and within studies that use different entropy measures, equipment, sampling frequencies or data collection durations. Methods and Findings The complexity of synthetic signals with known properties and standing CoP data was calculated using Approximate Entropy (ApEn), Sample Entropy (SampEn) and Recurrence Quantification Analysis Entropy (RQAEn). All signals were examined at varying sampling frequencies and with varying amounts of added noise. Additionally, an increment time series of the original CoP data was examined to remove long-range correlations. Of the three measures examined, ApEn was the least robust to sampling frequency and noise manipulations. Additionally, increased noise led to an increase in SampEn, but a decrease in RQAEn. Thus, noise can yield inconsistent results between the various entropy measures. Finally, the differences between the entropy measures were minimized in the increment CoP data, suggesting that long-range correlations should be removed from CoP data prior to calculating entropy. Conclusions The various algorithms typically used to quantify the complexity (entropy) of CoP may yield very different results, particularly when sampling frequency and noise are different. The results of this study are discussed within the context of the neural noise and loss of complexity hypotheses. PMID:21437281
NASA Technical Reports Server (NTRS)
Troudet, Terry; Merrill, Walter C.
1990-01-01
The ability of feed-forward neural network architectures to learn continuous valued mappings in the presence of noise was demonstrated in relation to parameter identification and real-time adaptive control applications. An error function was introduced to help optimize parameter values such as number of training iterations, observation time, sampling rate, and scaling of the control signal. The learning performance depended essentially on the degree of embodiment of the control law in the training data set and on the degree of uniformity of the probability distribution function of the data that are presented to the net during sequence. When a control law was corrupted by noise, the fluctuations of the training data biased the probability distribution function of the training data sequence. Only if the noise contamination is minimized and the degree of embodiment of the control law is maximized, can a neural net develop a good representation of the mapping and be used as a neurocontroller. A multilayer net was trained with back-error-propagation to control a cart-pole system for linear and nonlinear control laws in the presence of data processing noise and measurement noise. The neurocontroller exhibited noise-filtering properties and was found to operate more smoothly than the teacher in the presence of measurement noise.
Assessment of Type I Interferon Signaling in Pediatric Inflammatory Disease.
Rice, Gillian I; Melki, Isabelle; Frémond, Marie-Louise; Briggs, Tracy A; Rodero, Mathieu P; Kitabayashi, Naoki; Oojageer, Anthony; Bader-Meunier, Brigitte; Belot, Alexandre; Bodemer, Christine; Quartier, Pierre; Crow, Yanick J
2017-02-01
Increased type I interferon is considered relevant to the pathology of a number of monogenic and complex disorders spanning pediatric rheumatology, neurology, and dermatology. However, no test exists in routine clinical practice to identify enhanced interferon signaling, thus limiting the ability to diagnose and monitor treatment of these diseases. Here, we set out to investigate the use of an assay measuring the expression of a panel of interferon-stimulated genes (ISGs) in children affected by a range of inflammatory diseases. A cohort study was conducted between 2011 and 2016 at the University of Manchester, UK, and the Institut Imagine, Paris, France. RNA PAXgene blood samples and clinical data were collected from controls and symptomatic patients with a genetically confirmed or clinically well-defined inflammatory phenotype. The expression of six ISGs was measured by quantitative polymerase chain reaction, and the median fold change was used to calculate an interferon score (IS) for each subject compared to a previously derived panel of 29 controls (where +2 SD of the control data, an IS of >2.466, is considered as abnormal). Results were correlated with genetic and clinical data. Nine hundred ninety-two samples were analyzed from 630 individuals comprising symptomatic patients across 24 inflammatory genotypes/phenotypes, unaffected heterozygous carriers, and controls. A consistent upregulation of ISG expression was seen in 13 monogenic conditions (455 samples, 265 patients; median IS 10.73, interquartile range (IQR) 5.90-18.41), juvenile systemic lupus erythematosus (78 samples, 55 patients; median IS 10.60, IQR 3.99-17.27), and juvenile dermatomyositis (101 samples, 59 patients; median IS 9.02, IQR 2.51-21.73) compared to controls (78 samples, 65 subjects; median IS 0.688, IQR 0.427-1.196), heterozygous mutation carriers (89 samples, 76 subjects; median IS 0.862, IQR 0.493-1.942), and individuals with non-molecularly defined autoinflammation (89 samples, 69 patients; median IS 1.07, IQR 0.491-3.74). An assessment of six ISGs can be used to define a spectrum of inflammatory diseases related to enhanced type I interferon signaling. If future studies demonstrate that the IS is a reactive biomarker, this measure may prove useful both in the diagnosis and the assessment of treatment efficacy.
NASA Astrophysics Data System (ADS)
Yamada, Y.; Gouda, N.; Yano, T.; Sako, N.; Hatsutori, Y.; Tanaka, T.; Yamauchi, M.
We explain simulation tools in JASMINE project(JASMINE simulator). The JASMINE project stands at the stage where its basic design will be determined in a few years. Then it is very important to simulate the data stream generated by astrometric fields at JASMINE in order to support investigations of error budgets, sampling strategy, data compression, data analysis, scientific performances, etc. Of course, component simulations are needed, but total simulations which include all components from observation target to satellite system are also very important. We find that new software technologies, such as Object Oriented(OO) methodologies are ideal tools for the simulation system of JASMINE(the JASMINE simulator). The simulation system should include all objects in JASMINE such as observation techniques, models of instruments and bus design, orbit, data transfer, data analysis etc. in order to resolve all issues which can be expected beforehand and make it easy to cope with some unexpected problems which might occur during the mission of JASMINE. So, the JASMINE Simulator is designed as handling events such as photons from astronomical objects, control signals for devices, disturbances for satellite attitude, by instruments such as mirrors and detectors, successively. The simulator is also applied to the technical demonstration "Nano-JASMINE". The accuracy of ordinary sensor is not enough for initial phase attitude control. Mission instruments may be a good sensor for this purpose. The problem of attitude control in initial phase is a good example of this software because the problem is closely related to both mission instruments and satellite bus systems.
NASA Astrophysics Data System (ADS)
Yamada, Y.; Gouda, N.; Yano, T.; Kobayashi, Y.; Suganuma, M.; Tsujimoto, T.; Sako, N.; Hatsutori, Y.; Tanaka, T.
2006-08-01
We explain simulation tools in JASMINE project (JASMINE simulator). The JASMINE project stands at the stage where its basic design will be determined in a few years. Then it is very important to simulate the data stream generated by astrometric fields at JASMINE in order to support investigations of error budgets, sampling strategy, data compression, data analysis, scientific performances, etc. Of course, component simulations are needed, but total simulations which include all components from observation target to satellite system are also very important. We find that new software technologies, such as Object Oriented (OO) methodologies are ideal tools for the simulation system of JASMINE (the JASMINE simulator). The simulation system should include all objects in JASMINE such as observation techniques, models of instruments and bus design, orbit, data transfer, data analysis etc. in order to resolve all issues which can be expected beforehand and make it easy to cope with some unexpected problems which might occur during the mission of JASMINE. So, the JASMINE Simulator is designed as handling events such as photons from astronomical objects, control signals for devices, disturbances for satellite attitude, by instruments such as mirrors and detectors, successively. The simulator is also applied to the technical demonstration "Nano-JASMINE". The accuracy of ordinary sensor is not enough for initial phase attitude control. Mission instruments may be a good sensor for this purpose. The problem of attitude control in initial phase is a good example of this software because the problem is closely related to both mission instruments and satellite bus systems.
Wright, Peter R.; McMahon, Peter B.; Mueller, David K.; Clark, Melanie L.
2012-01-01
In June 2010, the U.S. Environmental Protection Agency installed two deep monitoring wells (MW01 and MW02) near Pavillion, Wyoming, to study groundwater quality. During April and May 2012, the U.S Geological Survey, in cooperation with the Wyoming Department of Environmental Quality, collected groundwater-quality data and quality-control data from monitoring well MW01 and, following well redevelopment, quality-control data for monitoring well MW02. Two groundwater-quality samples were collected from well MW01—one sample was collected after purging about 1.5 borehole volumes, and a second sample was collected after purging 3 borehole volumes. Both samples were collected and processed using methods designed to minimize atmospheric contamination or changes to water chemistry. Groundwater-quality samples were analyzed for field water-quality properties (water temperature, pH, specific conductance, dissolved oxygen, oxidation potential); inorganic constituents including naturally occurring radioactive compounds (radon, radium-226 and radium-228); organic constituents; dissolved gasses; stable isotopes of methane, water, and dissolved inorganic carbon; and environmental tracers (carbon-14, chlorofluorocarbons, sulfur hexafluoride, tritium, helium, neon, argon, krypton, xenon, and the ratio of helium-3 to helium-4). Quality-control sample results associated with well MW01 were evaluated to determine the extent to which environmental sample analytical results were affected by bias and to evaluate the variability inherent to sample collection and laboratory analyses. Field documentation, environmental data, and quality-control data for activities that occurred at the two monitoring wells during April and May 2012 are presented.
Autonomous support for microorganism research in space
NASA Astrophysics Data System (ADS)
Fleet, M. L.; Smith, J. D.; Klaus, D. M.; Luttges, M. W.
1993-02-01
A preliminary design for performing on orbit, autonomous research on microorganisms and cultured cells/tissues is presented. The payload is designed to be compatible with the COMercial Experiment Transporter (COMET), an orbiter middeck locker interface and with Space Station Freedom. Uplink/downlink capabilities and sample return through controlled reentry are available for all carriers. Autonomous testing activities are preprogrammed with in-flight reprogrammability. Sensors for monitoring temperature, pH, light, gravity levels, vibrations, and radiation are provided for environmental regulation and experimental data collection. Additional data acquisition includes optical density measurement, microscopy, video, and film photography. On-board data storage capabilities are provided. A fluid transfer mechanism is utilized for inoculation, sampling, and nutrient replenishment of experiment cultures. In addition to payload design, research opportunities are explored to illustrate hardware versatility and function. The project is defined to provide biological data pertinent to extended duration crewed space flight including crew health issues and development of a Controlled Ecological Life Support System (CELSS). In addition, opportunities are opened for investigations leading to commercial applications of space, such as pharmaceutical development, modeling of terrestrial diseases, and material processing.
NASA Astrophysics Data System (ADS)
Jenness, T.; Robson, E. I.; Stevens, J. A.
2010-01-01
Calibrated data for 143 flat-spectrum extragalactic radio sources are presented at a wavelength of 850μm covering a 5-yr period from 2000 April. The data, obtained at the James Clerk Maxwell Telescope using the Submillimetre Common-User Bolometer Array (SCUBA) camera in pointing mode, were analysed using an automated pipeline process based on the Observatory Reduction and Acquisition Control - Data Reduction (ORAC-DR) system. This paper describes the techniques used to analyse and calibrate the data, and presents the data base of results along with a representative sample of the better-sampled light curves. A re-analysis of previously published data from 1997 to 2000 is also presented. The combined catalogue, comprising 10493 flux density measurements, provides a unique and valuable resource for studies of extragalactic radio sources.
Shipboard Analytical Capabilities on the Renovated JOIDES Resolution, IODP Riserless Drilling Vessel
NASA Astrophysics Data System (ADS)
Blum, P.; Foster, P.; Houpt, D.; Bennight, C.; Brandt, L.; Cobine, T.; Crawford, W.; Fackler, D.; Fujine, K.; Hastedt, M.; Hornbacher, D.; Mateo, Z.; Moortgat, E.; Vasilyev, M.; Vasilyeva, Y.; Zeliadt, S.; Zhao, J.
2008-12-01
The JOIDES Resolution (JR) has conducted 121 scientific drilling expeditions during the Ocean Drilling Program (ODP) and the first phase of the Integrated Ocean Drilling Program (IODP) (1983-2006). The vessel and scientific systems have just completed an NSF-sponsored renovation (2005-2008). Shipboard analytical systems have been upgraded, within funding constraints imposed by market driven vessel conversion cost increases, to include: (1) enhanced shipboard analytical services including instruments and software for sampling and the capture of chemistry, physical properties, and geological data; (2) new data management capabilities built around a laboratory information management system (LIMS), digital asset management system, and web services; (3) operations data services with enhanced access to navigation and rig instrumentation data; and (4) a combination of commercial and home-made user applications for workflow- specific data extractions, generic and customized data reporting, and data visualization within a shipboard production environment. The instrumented data capture systems include a new set of core loggers for rapid and non-destructive acquisition of images and other physical properties data from drill cores. Line-scan imaging and natural gamma ray loggers capture data at unprecedented quality due to new and innovative designs. Many instruments used to characterize chemical compounds of rocks, sediments, and interstitial fluids were upgraded with the latest technology. The shipboard analytical environment features a new and innovative framework (DESCinfo) and application (DESClogik) for capturing descriptive and interpretive data from geological sub-domains such as sedimentology, petrology, paleontology, structural geology, stratigraphy, etc. This system fills a long-standing gap by providing a global database, controlled vocabularies and taxa name lists with version control, a highly configurable spreadsheet environment for data capture, and visualization of context data collected with the shipboard core loggers and other instruments.
Mammary candidiasis: molecular-based detection of Candida species in human milk samples.
Mutschlechner, W; Karall, D; Hartmann, C; Streiter, B; Baumgartner-Sigl, S; Orth-Höller, D; Lass-Flörl, C
2016-08-01
In this prospective and monocentric study, we investigated the performance of a commercialized real-time polymerase chain reaction (RT-PCR) test system for the specific detection of DNA from Candida albicans, C. dubliniensis, C. glabrata, C. krusei, C. lusitaniae, C. parapsilosis, and C. tropicalis in human milk samples of patients suspicious of mammary candidiasis. For this purpose, 43 breast-feeding women with characteristic symptoms of mammary candidiasis and 40 asymptomatic controls were enrolled. By culture, Candida spp. were detected in 8.8 % (4/46) and 9.3 % (4/43) of patient and control samples, respectively. Candida albicans (2/46), C. parapsilosis (1/46), and C. guilliermondii (1/46) were present in patient samples, and C. lusitaniae (3/43) and C. guilliermondii (1/43) were present in the controls. After RT-PCR was applied, Candida spp. were found to be present in 67.4 % (31/46) and 79.1 % (34/43) of patient and control samples investigated, respectively. PCR detection of C. albicans and C. parapsilosis revealed only a low sensitivity and specificity of 67.4 % and 41.9 %, respectively. Our data do not support the use of Candida RT-PCR for sensitive and specific diagnosis of mammary candidiasis.
Design of extensible meteorological data acquisition system based on FPGA
NASA Astrophysics Data System (ADS)
Zhang, Wen; Liu, Yin-hua; Zhang, Hui-jun; Li, Xiao-hui
2015-02-01
In order to compensate the tropospheric refraction error generated in the process of satellite navigation and positioning. Temperature, humidity and air pressure had to be used in concerned models to calculate the value of this error. While FPGA XC6SLX16 was used as the core processor, the integrated silicon pressure sensor MPX4115A and digital temperature-humidity sensor SHT75 are used as the basic meteorological parameter detection devices. The core processer was used to control the real-time sampling of ADC AD7608 and to acquire the serial output data of SHT75. The data was stored in the BRAM of XC6SLX16 and used to generate standard meteorological parameters in NEMA format. The whole design was based on Altium hardware platform and ISE software platform. The system was described in the VHDL language and schematic diagram to realize the correct detection of temperature, humidity, air pressure. The 8-channel synchronous sampling characteristics of AD7608 and programmable external resources of FPGA laid the foundation for the increasing of analog or digital meteorological element signal. The designed meteorological data acquisition system featured low cost, high performance, multiple expansions.
Postural control assessment in students with normal hearing and sensorineural hearing loss.
Melo, Renato de Souza; Lemos, Andrea; Macky, Carla Fabiana da Silva Toscano; Raposo, Maria Cristina Falcão; Ferraz, Karla Mônica
2015-01-01
Children with sensorineural hearing loss can present with instabilities in postural control, possibly as a consequence of hypoactivity of their vestibular system due to internal ear injury. To assess postural control stability in students with normal hearing (i.e., listeners) and with sensorineural hearing loss, and to compare data between groups, considering gender and age. This cross-sectional study evaluated the postural control of 96 students, 48 listeners and 48 with sensorineural hearing loss, aged between 7 and 18 years, of both genders, through the Balance Error Scoring Systems scale. This tool assesses postural control in two sensory conditions: stable surface and unstable surface. For statistical data analysis between groups, the Wilcoxon test for paired samples was used. Students with hearing loss showed more instability in postural control than those with normal hearing, with significant differences between groups (stable surface, unstable surface) (p<0.001). Students with sensorineural hearing loss showed greater instability in the postural control compared to normal hearing students of the same gender and age. Copyright © 2014 Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial. Published by Elsevier Editora Ltda. All rights reserved.
Purdue ionomics information management system. An integrated functional genomics platform.
Baxter, Ivan; Ouzzani, Mourad; Orcun, Seza; Kennedy, Brad; Jandhyala, Shrinivas S; Salt, David E
2007-02-01
The advent of high-throughput phenotyping technologies has created a deluge of information that is difficult to deal with without the appropriate data management tools. These data management tools should integrate defined workflow controls for genomic-scale data acquisition and validation, data storage and retrieval, and data analysis, indexed around the genomic information of the organism of interest. To maximize the impact of these large datasets, it is critical that they are rapidly disseminated to the broader research community, allowing open access for data mining and discovery. We describe here a system that incorporates such functionalities developed around the Purdue University high-throughput ionomics phenotyping platform. The Purdue Ionomics Information Management System (PiiMS) provides integrated workflow control, data storage, and analysis to facilitate high-throughput data acquisition, along with integrated tools for data search, retrieval, and visualization for hypothesis development. PiiMS is deployed as a World Wide Web-enabled system, allowing for integration of distributed workflow processes and open access to raw data for analysis by numerous laboratories. PiiMS currently contains data on shoot concentrations of P, Ca, K, Mg, Cu, Fe, Zn, Mn, Co, Ni, B, Se, Mo, Na, As, and Cd in over 60,000 shoot tissue samples of Arabidopsis (Arabidopsis thaliana), including ethyl methanesulfonate, fast-neutron and defined T-DNA mutants, and natural accession and populations of recombinant inbred lines from over 800 separate experiments, representing over 1,000,000 fully quantitative elemental concentrations. PiiMS is accessible at www.purdue.edu/dp/ionomics.
The Information Super Seaway: Networking the Seafloor for Interactive Scientific Discovery
NASA Astrophysics Data System (ADS)
Daly, K. L.; Isern, A. R.
2006-05-01
Ship-based expeditionary research and satellite observations have provided basic descriptions of ocean processes and their interactions with terrestrial and atmospheric systems. Many critical processes, however, occur at temporal and spatial scales that cannot be effectively sampled or studied with these traditional techniques. Ship-based studies are particularly limited in their ability to investigate the onset and immediate aftermath of episodic events and non-linear processes. Enabled by technological advances and made timely by societal need, a wide range of ocean and earth observing systems are being planned, proposed, deployed and operated within the U.S. These systems will utilize real-time datasets for event detection and adaptive sampling, well-sampled spatial and temporal contexts for limited duration or process-study experiments, and sustained observations to observe long-term trends and capture rare episodic events. Recent developments in sensor technology, cyberinfrastructure, and modeling capabilities will enable scientists to consider an entirely new set of interdisciplinary science questions. In response to the need for long term in situ oceanographic data, the U.S. National Science Foundation has established the Ocean Research Interactive Observatory Networks (ORION) Program within which the Ocean Observatories Initiative (OOI) will provide the essential infrastructure to address high priority science questions outlined in the OOI Science Plan. This infrastructure will utilize electro-optical cables and moored buoys to enable real-time, high bandwidth transmissions of scientific data and images from key scientific sites in the coastal and open ocean. The OOI is an integrated observatory with three elements: 1) a regional cabled network consisting of interconnected sites on the seafloor spanning several geological and oceanographic features and processes, 2) relocatable deep-sea buoys that could also be deployed in harsh environments such as the Southern Ocean, and 3) new construction or enhancements to existing facilities leading to an expanded network of coastal observatories. The ORION Program will coordinate the science driving the construction of this research observing network as well as the operation and maintenance of the infrastructure; development of instrumentation and mobile platforms and their incorporation into the observatory network; and planning, coordination, and implementation of educational and public outreach activities. A critical integrating element of the seafloor observatory network will be a robust cyberinfrastructure system that can collect large volumes of heterogeneous data. This system is being developed to collect, manage, archive and distribute data; have mechanisms and protocols for rapid data transmission; have protocols for two-way communication with sensors and dynamic control of sensor networks; have access to remote computing resources for processing and visualization of data; have software tools for analysis of multidisciplinary, spatially extended, intermittent datasets; have knowledge representation software to ensure that these data are easily accessible and effortlessly shared across disciplines; have integrity between communications and control systems and data management and archiving systems; and have automated data quality control. The ORION Program will be the most complex initiative that ocean scientists have undertaken within the U.S. and will revolutionize the way that oceanographers study the sea.
Rudnick, Paul A.; Clauser, Karl R.; Kilpatrick, Lisa E.; Tchekhovskoi, Dmitrii V.; Neta, Pedatsur; Blonder, Nikša; Billheimer, Dean D.; Blackman, Ronald K.; Bunk, David M.; Cardasis, Helene L.; Ham, Amy-Joan L.; Jaffe, Jacob D.; Kinsinger, Christopher R.; Mesri, Mehdi; Neubert, Thomas A.; Schilling, Birgit; Tabb, David L.; Tegeler, Tony J.; Vega-Montoto, Lorenzo; Variyath, Asokan Mulayath; Wang, Mu; Wang, Pei; Whiteaker, Jeffrey R.; Zimmerman, Lisa J.; Carr, Steven A.; Fisher, Susan J.; Gibson, Bradford W.; Paulovich, Amanda G.; Regnier, Fred E.; Rodriguez, Henry; Spiegelman, Cliff; Tempst, Paul; Liebler, Daniel C.; Stein, Stephen E.
2010-01-01
A major unmet need in LC-MS/MS-based proteomics analyses is a set of tools for quantitative assessment of system performance and evaluation of technical variability. Here we describe 46 system performance metrics for monitoring chromatographic performance, electrospray source stability, MS1 and MS2 signals, dynamic sampling of ions for MS/MS, and peptide identification. Applied to data sets from replicate LC-MS/MS analyses, these metrics displayed consistent, reasonable responses to controlled perturbations. The metrics typically displayed variations less than 10% and thus can reveal even subtle differences in performance of system components. Analyses of data from interlaboratory studies conducted under a common standard operating procedure identified outlier data and provided clues to specific causes. Moreover, interlaboratory variation reflected by the metrics indicates which system components vary the most between laboratories. Application of these metrics enables rational, quantitative quality assessment for proteomics and other LC-MS/MS analytical applications. PMID:19837981
Mischnik, Alexander; Mieth, Markus; Busch, Cornelius J; Hofer, Stefan; Zimmermann, Stefan
2012-08-01
Automation of plate streaking is ongoing in clinical microbiological laboratories, but evaluation for routine use is mostly open. In the present study, the recovery of microorganisms from the Previ Isola system plated polyurethane (PU) swab samples is compared to manually plated control viscose swab samples from wounds according to the CLSI procedure M40-A (quality control of microbiological transport systems). One hundred twelve paired samples (224 swabs) were analyzed. In 80/112 samples (71%), concordant culture results were obtained with the two methods. In 32/112 samples (29%), CFU recovery of microorganisms from the two methods was discordant. In 24 (75%) of the 32 paired samples with a discordant result, Previ Isola plated PU swabs were superior. In 8 (25%) of the 32 paired samples with a discordant result, control viscose swabs were superior. The quality of colony growth on culture media for further investigations was superior with Previ Isola inoculated plates compared to manual plating techniques. Gram stain results were concordant between the two methods in 62/112 samples (55%). In 50/112 samples (45%), the results of Gram staining were discordant between the two methods. In 34 (68%) of the 50 paired samples with discordant results, Gram staining of PU swabs was superior to that of control viscose swabs. In 16 (32%) of the 50 paired samples, Gram staining of control viscose swabs was superior to that of PU swabs. We report the first clinical evaluation of Previ Isola automated specimen inoculation for wound swab samples. This study suggests that use of an automated specimen inoculation system has good results with regard to CFU recovery, quality of Gram staining, and accuracy of diagnosis.
Mieth, Markus; Busch, Cornelius J.; Hofer, Stefan; Zimmermann, Stefan
2012-01-01
Automation of plate streaking is ongoing in clinical microbiological laboratories, but evaluation for routine use is mostly open. In the present study, the recovery of microorganisms from the Previ Isola system plated polyurethane (PU) swab samples is compared to manually plated control viscose swab samples from wounds according to the CLSI procedure M40-A (quality control of microbiological transport systems). One hundred twelve paired samples (224 swabs) were analyzed. In 80/112 samples (71%), concordant culture results were obtained with the two methods. In 32/112 samples (29%), CFU recovery of microorganisms from the two methods was discordant. In 24 (75%) of the 32 paired samples with a discordant result, Previ Isola plated PU swabs were superior. In 8 (25%) of the 32 paired samples with a discordant result, control viscose swabs were superior. The quality of colony growth on culture media for further investigations was superior with Previ Isola inoculated plates compared to manual plating techniques. Gram stain results were concordant between the two methods in 62/112 samples (55%). In 50/112 samples (45%), the results of Gram staining were discordant between the two methods. In 34 (68%) of the 50 paired samples with discordant results, Gram staining of PU swabs was superior to that of control viscose swabs. In 16 (32%) of the 50 paired samples, Gram staining of control viscose swabs was superior to that of PU swabs. We report the first clinical evaluation of Previ Isola automated specimen inoculation for wound swab samples. This study suggests that use of an automated specimen inoculation system has good results with regard to CFU recovery, quality of Gram staining, and accuracy of diagnosis. PMID:22692745
Evaluation of a novel chemical sensor system to detect clinical mastitis in bovine milk.
Mottram, Toby; Rudnitskaya, Alisa; Legin, Andrey; Fitzpatrick, Julie L; Eckersall, P David
2007-05-15
Automatic detection of clinical mastitis is an essential part of high performance and robotic milking. Currently available technology (conductivity monitoring) is unable to achieve acceptable specificity or sensitivity of detection of clinical mastitis or other clinical diseases. Arrays of sensors with high cross-sensitivity have been successfully applied for recognition and quantitative analysis of other multicomponent liquids. An experiment was conducted to determine whether a multisensor system ("electronic tongue") based on an array of chemical sensors and suitable data processing could be used to discriminate between milk secretions from infected and healthy glands. Measurements were made with a multisensor system of milk samples from two different farms in two experiments. A total of 67 samples of milk from both mastitic and healthy glands were in two sets. It was demonstrated that the multisensor system could distinguish between control and clinically mastitic milk samples (p=0.05). The sensitivity and specificity of the sensor system (93 and 96% correspondingly) showed an improvement over conductivity (56 and 82% correspondingly). The multisensor system offers a novel method of improving mastitis detection.
ALMA Correlator Real-Time Data Processor
NASA Astrophysics Data System (ADS)
Pisano, J.; Amestica, R.; Perez, J.
2005-10-01
The design of a real-time Linux application utilizing Real-Time Application Interface (RTAI) to process real-time data from the radio astronomy correlator for the Atacama Large Millimeter Array (ALMA) is described. The correlator is a custom-built digital signal processor which computes the cross-correlation function of two digitized signal streams. ALMA will have 64 antennas with 2080 signal streams each with a sample rate of 4 giga-samples per second. The correlator's aggregate data output will be 1 gigabyte per second. The software is defined by hard deadlines with high input and processing data rates, while requiring interfaces to non real-time external computers. The designed computer system - the Correlator Data Processor or CDP, consists of a cluster of 17 SMP computers, 16 of which are compute nodes plus a master controller node all running real-time Linux kernels. Each compute node uses an RTAI kernel module to interface to a 32-bit parallel interface which accepts raw data at 64 megabytes per second in 1 megabyte chunks every 16 milliseconds. These data are transferred to tasks running on multiple CPUs in hard real-time using RTAI's LXRT facility to perform quantization corrections, data windowing, FFTs, and phase corrections for a processing rate of approximately 1 GFLOPS. Highly accurate timing signals are distributed to all seventeen computer nodes in order to synchronize them to other time-dependent devices in the observatory array. RTAI kernel tasks interface to the timing signals providing sub-millisecond timing resolution. The CDP interfaces, via the master node, to other computer systems on an external intra-net for command and control, data storage, and further data (image) processing. The master node accesses these external systems utilizing ALMA Common Software (ACS), a CORBA-based client-server software infrastructure providing logging, monitoring, data delivery, and intra-computer function invocation. The software is being developed in tandem with the correlator hardware which presents software engineering challenges as the hardware evolves. The current status of this project and future goals are also presented.
SAIL--a software system for sample and phenotype availability across biobanks and cohorts.
Gostev, Mikhail; Fernandez-Banet, Julio; Rung, Johan; Dietrich, Joern; Prokopenko, Inga; Ripatti, Samuli; McCarthy, Mark I; Brazma, Alvis; Krestyaninova, Maria
2011-02-15
The Sample avAILability system-SAIL-is a web based application for searching, browsing and annotating biological sample collections or biobank entries. By providing individual-level information on the availability of specific data types (phenotypes, genetic or genomic data) and samples within a collection, rather than the actual measurement data, resource integration can be facilitated. A flexible data structure enables the collection owners to provide descriptive information on their samples using existing or custom vocabularies. Users can query for the available samples by various parameters combining them via logical expressions. The system can be scaled to hold data from millions of samples with thousands of variables. SAIL is available under Aferro-GPL open source license: https://github.com/sail.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Daley, P F
The overall objective of this project is the continued development, installation, and testing of continuous water sampling and analysis technologies for application to on-site monitoring of groundwater treatment systems and remediation sites. In a previous project, an on-line analytical system (OLAS) for multistream water sampling was installed at the Fort Ord Operable Unit 2 Groundwater Treatment System, with the objective of developing a simplified analytical method for detection of Compounds of Concern at that plant, and continuous sampling of up to twelve locations in the treatment system, from raw influent waters to treated effluent. Earlier implementations of the water samplingmore » and processing system (Analytical Sampling and Analysis Platform, A A+RT, Milpitas, CA) depended on off-line integrators that produced paper plots of chromatograms, and sent summary tables to a host computer for archiving. We developed a basic LabVIEW (National Instruments, Inc., Austin, TX) based gas chromatography control and data acquisition system that was the foundation for further development and integration with the ASAP system. Advantages of this integration include electronic archiving of all raw chromatographic data, and a flexible programming environment to support development of improved ASAP operation and automated reporting. The initial goals of integrating the preexisting LabVIEW chromatography control system with the ASAP, and demonstration of a simplified, site-specific analytical method were successfully achieved. However, although the principal objective of this system was assembly of an analytical system that would allow plant operators an up-to-the-minute view of the plant's performance, several obstacles remained. Data reduction with the base LabVIEW system was limited to peak detection and simple tabular output, patterned after commercial chromatography integrators, with compound retention times and peak areas. Preparation of calibration curves, method detection limit estimates and trend plotting were performed with spreadsheets and statistics software. Moreover, the analytical method developed was very limited in compound coverage, and unable to closely mirror the standard analytical methods promulgated by the EPA. To address these deficiencies, during this award the original equipment was operated at the OU 2-GTS to further evaluate the use of columns, commercial standard blends and other components to broaden the compound coverage of the chromatography system. A second-generation ASAP was designed and built to replace the original system at the OU 2-GTS, and include provision for introduction of internal standard compounds and surrogates into each sample analyzed. An enhanced, LabVIEW based chromatogram analysis application was written, that manages and archives chemical standards information, and provides a basis for NIST traceability for all analyses. Within this same package, all compound calibration response curves are managed, and different report formats were incorporated, that simplify trend analysis. Test results focus on operation of the original system at the OU 1 Integrated Chemical and Flow Monitoring System, at the OU 1 Fire Drill Area remediation site.« less
NASA Astrophysics Data System (ADS)
Bou-Fakhreddine, Bassam; Mougharbel, Imad; Faye, Alain; Abou Chakra, Sara; Pollet, Yann
2018-03-01
Accurate daily river flow forecast is essential in many applications of water resources such as hydropower operation, agricultural planning and flood control. This paper presents a forecasting approach to deal with a newly addressed situation where hydrological data exist for a period longer than that of meteorological data (measurements asymmetry). In fact, one of the potential solutions to resolve measurements asymmetry issue is data re-sampling. It is a matter of either considering only the hydrological data or the balanced part of the hydro-meteorological data set during the forecasting process. However, the main disadvantage is that we may lose potentially relevant information from the left-out data. In this research, the key output is a Two-Phase Constructive Fuzzy inference hybrid model that is implemented over the non re-sampled data. The introduced modeling approach must be capable of exploiting the available data efficiently with higher prediction efficiency relative to Constructive Fuzzy model trained over re-sampled data set. The study was applied to Litani River in the Bekaa Valley - Lebanon by using 4 years of rainfall and 24 years of river flow daily measurements. A Constructive Fuzzy System Model (C-FSM) and a Two-Phase Constructive Fuzzy System Model (TPC-FSM) are trained. Upon validating, the second model has shown a primarily competitive performance and accuracy with the ability to preserve a higher day-to-day variability for 1, 3 and 6 days ahead. In fact, for the longest lead period, the C-FSM and TPC-FSM were able of explaining respectively 84.6% and 86.5% of the actual river flow variation. Overall, the results indicate that TPC-FSM model has provided a better tool to capture extreme flows in the process of streamflow prediction.
SUBOPT: A CAD program for suboptimal linear regulators
NASA Technical Reports Server (NTRS)
Fleming, P. J.
1985-01-01
An interactive software package which provides design solutions for both standard linear quadratic regulator (LQR) and suboptimal linear regulator problems is described. Intended for time-invariant continuous systems, the package is easily modified to include sampled-data systems. LQR designs are obtained by established techniques while the large class of suboptimal problems containing controller and/or performance index options is solved using a robust gradient minimization technique. Numerical examples demonstrate features of the package and recent developments are described.
Programmable temperature control system for biological materials
NASA Technical Reports Server (NTRS)
Anselmo, V. J.; Harrison, R. G.; Rinfret, A. P.
1982-01-01
A system was constructed which allows programmable temperature-time control for a 5 cu cm sample volume of arbitrary biological material. The system also measures the parameters necessary for the determination of the sample volume specific heat and thermal conductivity as a function of temperature, and provides a detailed measurement of the temperature during phase change and a means of calculating the heat of the phase change. Steady-state and dynamic temperature control is obtained by supplying heat to the sample volume through resistive elements constructed as an integral part of the sample container. For cooling purposes, this container is totally immersed into a cold heat sink. Using a mixture of dry ice and alcohol at 79 C, the sample volume can be controlled from +40 to -60 C at rates from steady state to + or - 65 C/min. Steady-state temperature precision is better than 0.2 C, while the dynamic capability depends on the temperature rate of change as well as the mass of both the sample and the container.
Automated site characterization for robotic sample acquisition systems
NASA Astrophysics Data System (ADS)
Scholl, Marija S.; Eberlein, Susan J.
1993-04-01
A mobile, semiautonomous vehicle with multiple sensors and on-board intelligence is proposed for performing preliminary scientific investigations on extraterrestrial bodies prior to human exploration. Two technologies, a hybrid optical-digital computer system based on optical correlator technology and an image and instrument data analysis system, provide complementary capabilities that might be part of an instrument package for an intelligent robotic vehicle. The hybrid digital-optical vision system could perform real-time image classification tasks using an optical correlator with programmable matched filters under control of a digital microcomputer. The data analysis system would analyze visible and multiband imagery to extract mineral composition and textural information for geologic characterization. Together these technologies would support the site characterization needs of a robotic vehicle for both navigational and scientific purposes.
Wasserman, Ira; Stack, Steven
2011-02-01
Previous work on Russian roulette has focused on data from large cities. It is unclear if the epidemiological patterns based on large cities will replicate for the nation as a whole, and if the influence of minority status will be moderated by urban context. The present investigation fills these gaps by providing descriptive epidemiological data on Russian roulette for 17 states, and testing a hypothesis on urbanism as a moderator of the race-Russian roulette relationship. Data were taken from the National Violent Death Reporting System (2003-2006). They refer to 71 Russian roulette cases and a matched control group of 284 males who committed suicide by a gunshot wound to the head. Russian roulette suicides were more apt to be of minority status, younger, had a lower incidence of mental health problems, and were more likely to be utilizing alcohol than the controls. Differentiating the sample into larger and smaller urban areas, it was found that the risk of Russian roulette for African Americans was higher in larger urban areas. Epidemiological patterns in previous research on large city samples are largely replicated. The moderating influence of urban context is related to differential opportunity structures for risk-taking behavior. © 2011 The American Association of Suicidology.
Saulnier, George E; Castro, Janna C; Cook, Curtiss B
2014-05-01
Glucose control can be problematic in critically ill patients. We evaluated the impact of statistical transformation on interpretation of intensive care unit inpatient glucose control data. Point-of-care blood glucose (POC-BG) data derived from patients in the intensive care unit for 2011 was obtained. Box-Cox transformation of POC-BG measurements was performed, and distribution of data was determined before and after transformation. Different data subsets were used to establish statistical upper and lower control limits. Exponentially weighted moving average (EWMA) control charts constructed from April, October, and November data determined whether out-of-control events could be identified differently in transformed versus nontransformed data. A total of 8679 POC-BG values were analyzed. POC-BG distributions in nontransformed data were skewed but approached normality after transformation. EWMA control charts revealed differences in projected detection of out-of-control events. In April, an out-of-control process resulting in the lower control limit being exceeded was identified at sample 116 in nontransformed data but not in transformed data. October transformed data detected an out-of-control process exceeding the upper control limit at sample 27 that was not detected in nontransformed data. Nontransformed November results remained in control, but transformation identified an out-of-control event less than 10 samples into the observation period. Using statistical methods to assess population-based glucose control in the intensive care unit could alter conclusions about the effectiveness of care processes for managing hyperglycemia. Further study is required to determine whether transformed versus nontransformed data change clinical decisions about the interpretation of care or intervention results. © 2014 Diabetes Technology Society.
Saulnier, George E.; Castro, Janna C.
2014-01-01
Glucose control can be problematic in critically ill patients. We evaluated the impact of statistical transformation on interpretation of intensive care unit inpatient glucose control data. Point-of-care blood glucose (POC-BG) data derived from patients in the intensive care unit for 2011 was obtained. Box–Cox transformation of POC-BG measurements was performed, and distribution of data was determined before and after transformation. Different data subsets were used to establish statistical upper and lower control limits. Exponentially weighted moving average (EWMA) control charts constructed from April, October, and November data determined whether out-of-control events could be identified differently in transformed versus nontransformed data. A total of 8679 POC-BG values were analyzed. POC-BG distributions in nontransformed data were skewed but approached normality after transformation. EWMA control charts revealed differences in projected detection of out-of-control events. In April, an out-of-control process resulting in the lower control limit being exceeded was identified at sample 116 in nontransformed data but not in transformed data. October transformed data detected an out-of-control process exceeding the upper control limit at sample 27 that was not detected in nontransformed data. Nontransformed November results remained in control, but transformation identified an out-of-control event less than 10 samples into the observation period. Using statistical methods to assess population-based glucose control in the intensive care unit could alter conclusions about the effectiveness of care processes for managing hyperglycemia. Further study is required to determine whether transformed versus nontransformed data change clinical decisions about the interpretation of care or intervention results. PMID:24876620
Sampled Data Adaptive Digital Computer Control of Surface Ship Maneuvers
1976-06-01
0.53 feet. Systems fcr which fuel considerations are not a motivating 157 factor lay te designed without this part of the control law ta allow finer...COXXXQXxaQXQ«^2Q£>’^ o>- —,>->>>ozor X < a. Ps4 <i i— « aC _J o < a o-*»-» ujOO • •>- o • •oo«mo z o «j II II ** » < ii ii -^ -* -,-^a:- i—— * O.-IUJ
Dairy-cattle health in Gyeongnam, Korea.
Jong, S K; Gong, S K; Chung, H K; Dae, S H
2001-12-03
An animal-health monitoring system in the Gyeongnam area was started in 1997 to develop statistically valid data for use in estimating disease frequencies in dairy cattle, and the associated costs. The objectives of this study were to: (1) describe what was done to implement and maintain the system in Gyeongnam; (2) present selected disease frequencies; (3) discuss the epidemiological consideration of what was done and implications for results obtained. Veterinary medical officers (VMOs), comprising professors and graduate students from Gyeongsang National University, faculty of Gyeongnam Livestock Promotion Institute and clinic veterinarians, served as data collectors. After training on current disease and management problems of dairy cattle, interview techniques, sampling methods and data-collection instruments, the VMOs participated in selection of the sample herds and data gathering. Forty (n=40) of 167 dairy herds were selected randomly using a computer-generated list of random numbers and the VMOs visited farms once in a month for 12 months to collect data about management, disease, inventory, production, preventive treatment, financial and other relevant data. Strict data-quality control devices were used. Specific feed-back was developed for the producers and data collectors. The six disorders found most frequently in cows (from the highest to the lowest) were breeding problems, clinical mastitis, birth problems, gastrointestinal problems, metabolic problems and lameness. In young stock, respiratory, multiple system, breeding and gastrointestinal problems were predominant, whereas in calves, gastrointestinal, respiratory and integumental problems predominated.
The 1985 Army Experience Survey: Tabular Descriptions of First-Term Separatees. Volume 1
1986-01-01
Assistance. Survey data were processed through survey receipt control and sample management systems . Data were also keyed, edited, coded, and weighted. The...268-269 047A R135 AGE OF OLDEST CHILD. ..................... .. ..... 270-271 048 R136 REGION OF RESIDENCE WHEN YOU JOINED ARMY...100.0 STATISTIC VALUE D.F. PROD. CHISQUARE APPROX. 65.725 10 0.0000 272 R136 -- REGION OF RESIDENCE WHEN YOU JOINED ARMY RECODED - WHAT STATE WERE
The 1985 Army Experience Survey: Tabular Descriptions of Mid-Career Separatees. Volume 2
1986-01-01
Survey data were processed through survey receipt control and sample management systems . Data were also keyed, edited, coded, and weighted. The coding...270-271 048 R136 REGION OF RESIDENCE WHEN YOU JOINED ARMY ................. . 272-273 049 E137 # TERMS OF ACTIVE ENLISTMENT...STATISTIC VALUE D.F. PROB. CHISQUARE APPROX. 4.449 5 0.4868 I 272 R136 -- REGION OF RESIDENCE WHEN YOU JOINED ARMY RECODED - WHAT STATE WERE YOU LIVING IN
Smart Sensor Node Development, Testing and Implementation for Rocket Propulsion Systems
NASA Technical Reports Server (NTRS)
Mengers, Timothy R.; Shipley, John; Merrill, Richard; Eggett, Leon; Johnson, Mont; Morris, Jonathan; Figueroa, Fernando; Schmalzel, John; Turowski, Mark P.
2007-01-01
Successful design and implementation of an Integrated System Health Management (ISHM) approach for rocket propulsion systems requires the capability improve the reliability of complex systems by detecting and diagnosing problems. One of the critical elements in the ISHM is an intelligent sensor node for data acquisition that meets specific requirements for rocket motor testing including accuracy, sample rate and size/weight. Traditional data acquisition systems are calibrated in a controlled environment and guaranteed to perform bounded by their tested conditions. In a real world ISHM system, the data acquisition and signal conditioning needs to function in an uncontrolled environment. Development and testing of this sensor node focuses on a design with the ability to self check in order to extend calibration times, report internal faults and drifts and notify the overall system when the data acquisition is not performing as it should. All of this will be designed within a system that is flexible, requiring little re-design to be deployed on a wide variety of systems. Progress in this design and initial testing of prototype units will be reported.
Development and Flight Testing of an Adaptive Vehicle Health-Monitoring Architecture
NASA Technical Reports Server (NTRS)
Woodard, Stanley E.; Coffey, Neil C.; Gonzalez, Guillermo A.; Taylor, B. Douglas; Brett, Rube R.; Woodman, Keith L.; Weathered, Brenton W.; Rollins, Courtney H.
2002-01-01
On going development and testing of an adaptable vehicle health-monitoring architecture is presented. The architecture is being developed for a fleet of vehicles. It has three operational levels: one or more remote data acquisition units located throughout the vehicle; a command and control unit located within the vehicle, and, a terminal collection unit to collect analysis results from all vehicles. Each level is capable of performing autonomous analysis with a trained expert system. The expert system is parameterized, which makes it adaptable to be trained to both a user's subject reasoning and existing quantitative analytic tools. Communication between all levels is done with wireless radio frequency interfaces. The remote data acquisition unit has an eight channel programmable digital interface that allows the user discretion for choosing type of sensors; number of sensors, sensor sampling rate and sampling duration for each sensor. The architecture provides framework for a tributary analysis. All measurements at the lowest operational level are reduced to provide analysis results necessary to gauge changes from established baselines. These are then collected at the next level to identify any global trends or common features from the prior level. This process is repeated until the results are reduced at the highest operational level. In the framework, only analysis results are forwarded to the next level to reduce telemetry congestion. The system's remote data acquisition hardware and non-analysis software have been flight tested on the NASA Langley B757's main landing gear. The flight tests were performed to validate the following: the wireless radio frequency communication capabilities of the system, the hardware design, command and control; software operation and, data acquisition, storage and retrieval.
GPUbased, Microsecond Latency, HectoChannel MIMO Feedback Control of Magnetically Confined Plasmas
NASA Astrophysics Data System (ADS)
Rath, Nikolaus
Feedback control has become a crucial tool in the research on magnetic confinement of plasmas for achieving controlled nuclear fusion. This thesis presents a novel plasma feedback control system that, for the first time, employs a Graphics Processing Unit (GPU) for microsecond-latency, real-time control computations. This novel application area for GPU computing is opened up by a new system architecture that is optimized for low-latency computations on less than kilobyte sized data samples as they occur in typical plasma control algorithms. In contrast to traditional GPU computing approaches that target complex, high-throughput computations with massive amounts of data, the architecture presented in this thesis uses the GPU as the primary processing unit rather than as an auxiliary of the CPU, and data is transferred from A-D/D-A converters directly into GPU memory using peer-to-peer PCI Express transfers. The described design has been implemented in a new, GPU-based control system for the High-Beta Tokamak - Extended Pulse (HBT-EP) device. The system is built from commodity hardware and uses an NVIDIA GeForce GPU and D-TACQ A-D/D-A converters providing a total of 96 input and 64 output channels. The system is able to run with sampling periods down to 4 μs and latencies down to 8 μs. The GPU provides a total processing power of 1.5 x 1012 floating point operations per second. To illustrate the performance and versatility of both the general architecture and concrete implementation, a new control algorithm has been developed. The algorithm is designed for the control of multiple rotating magnetic perturbations in situations where the plasma equilibrium is not known exactly and features an adaptive system model: instead of requiring the rotation frequencies and growth rates embedded in the system model to be set a priori, the adaptive algorithm derives these parameters from the evolution of the perturbation amplitudes themselves. This results in non-linear control computations with high computational demands, but is handled easily by the GPU based system. Both digital processing latency and an arbitrary multi-pole response of amplifiers and control coils is fully taken into account for the generation of control signals. To separate sensor signals into perturbed and equilibrium components without knowledge of the equilibrium fields, a new separation method based on biorthogonal decomposition is introduced and used to derive a filter that performs the separation in real-time. The control algorithm has been implemented and tested on the new, GPU-based feedback control system of the HBT-EP tokamak. In this instance, the algorithm was set up to control four rotating n = 1 perturbations at different poloidal angles. The perturbations were treated as coupled in frequency but independent in amplitude and phase, so that the system effectively controls a helical n = 1 perturbation with unknown poloidal spectrum. Depending on the plasma's edge safety factor and rotation frequency, the control system is shown to be able to suppress the amplitude of the dominant 8 kHz mode by up to 60% or amplify the saturated amplitude by a factor of up to two. Intermediate feedback phases combine suppression and amplification with a speed up or slow down of the mode rotation frequency. Increasing feedback gain results in the excitation of an additional, slowly rotating 1.4 kHz mode without further effects on the 8 kHz mode. The feedback performance is found to exceed previous results obtained with an FPGA- and Kalman-filter based control system without requiring any tuning of system model parameters. Experimental results are compared with simulations based on a combination of the Boozer surface current model and the Fitzpatrick-Aydemir model. Within the subset of phenomena that can be represented by the model as well as determined experimentally, qualitative agreement is found.
Amelogenin test: From forensics to quality control in clinical and biochemical genomics.
Francès, F; Portolés, O; González, J I; Coltell, O; Verdú, F; Castelló, A; Corella, D
2007-01-01
The increasing number of samples from the biomedical genetic studies and the number of centers participating in the same involves increasing risk of mistakes in the different sample handling stages. We have evaluated the usefulness of the amelogenin test for quality control in sample identification. Amelogenin test (frequently used in forensics) was undertaken on 1224 individuals participating in a biomedical study. Concordance between referred sex in the database and amelogenin test was estimated. Additional sex-error genetic detecting systems were developed. The overall concordance rate was 99.84% (1222/1224). Two samples showed a female amelogenin test outcome, being codified as males in the database. The first, after checking sex-specific biochemical and clinical profile data was found to be due to a codification error in the database. In the second, after checking the database, no apparent error was discovered because a correct male profile was found. False negatives in amelogenin male sex determination were discarded by additional tests, and feminine sex was confirmed. A sample labeling error was revealed after a new DNA extraction. The amelogenin test is a useful quality control tool for detecting sex-identification errors in large genomic studies, and can contribute to increase its validity.
Supervisory Control and Data Acquisition System | Energy Systems
Integration Facility | NREL Supervisory Control and Data Acquisition System Supervisory Control supervisory control and data acquisition (SCADA) system monitors and controls safety systems and gathers real Energy Systems Integration Facility control room. The Energy Systems Integration Facility's SCADA system
Decreased maternal plasma apelin concentrations in preeclampsia.
Bortoff, Katherine D; Qiu, Chunfang; Runyon, Scott; Williams, Michelle A; Maitra, Rangan
2012-01-01
Preeclampsia is a hypertensive disorder that complicates 3-7% of pregnancies. The development of preeclampsia has not been completely elucidated and current therapies are not broadly efficacious. The apelinergic system appears to be involved in hypertensive disorders and experimental studies indicate a role of this system in preeclampsia. Thus, an epidemiological evaluation of apelin protein concentration in plasma was conducted in case-control study of pregnant women. Data and maternal plasma samples were collected from pregnant women with confirmed preeclampsia (n = 76) or normotensive controls (n = 79). Concentrations of apelin peptides were blindly measured using enzyme-linked immunosorbent assay. Data were subjected to statistical analyses. Plasma apelin concentrations, measured at delivery, were lower in preeclampsia cases compared with controls (mean ± standard deviation: 0.66 ± 0.29 vs. 0.78 ± 0.31 ng/mL, p = 0.02). After controlling for confounding by maternal age, smoking status, and pre-pregnancy body mass index, odds of preeclampsia were 48% lower for women with high versus low plasma apelin (≥0.73 vs. <0.73 ng/mL) concentrations. Reduced circulating apelin peptides may be associated with preeclampsia. The apelinergic system should be further investigated to elucidate its role in preclampsia and other hypertensive maternal disorders.
NASA Astrophysics Data System (ADS)
Lehnert, K. A.; Goldstein, S. L.; Vinayagamoorthy, S.; Lenhardt, W. C.
2005-12-01
Data on samples represent a primary foundation of Geoscience research across disciplines, ranging from the study of climate change, to biogeochemical cycles, to mantle and continental dynamics and are key to our knowledge of the Earth's dynamical systems and evolution. Different data types are generated for individual samples by different research groups, published in different papers, and stored in different databases on a global scale. The utility of these data is critically dependent on their integration. Such integration can be achieved within a Geoscience Cyberinfrastructure, but requires unambiguous identification of samples. Currently, naming of samples is arbitrary and inconsistent and therefore severely limits our ability to share, link, and integrate sample-based data. Major problems include name duplication, and changing of names as a sample is passed along over many years to different investigators. SESAR, the System for Earth Sample Registration (http://www.geosamples.org), addresses this problem by building a registry that generates and administers globally unique identifiers for Geoscience samples: the International Geo Sample Number (IGSN). Implementation of the IGSN in data publication and digital data management will dramatically advance interoperability among information systems for sample-based data, opening an extensive range of new opportunities for discovery and for interdisciplinary approaches in research. The IGSN will also facilitate the ability of investigators to build on previously collected data on samples as new measurements are made or new techniques are developed. With potentially broad application to all types of Geoscience samples, SESAR is global in scope. It is a web-based system that can be easily accessed by individual users through an interactive web interface and by distributed client systems via standard web services. Samples can be registered individually or in batches and at various levels of granularity from entire cores or dredges or sample suites to individual samples to sub-samples such as splits and separates. Relationships between `parent' and `child' samples are tracked. The system generates bar codes that users can download as images for labeling purposes. SESAR released a beta version of the registry in April 2005 that allows users to register a limited range of sample types. Identifiers generated by the beta version will remain valid when SESAR moves into its operational stage. Since then more than 3700 samples have been registered in SESAR. Registration of samples at a central clearinghouse will automatically build a global catalog of Geoscience samples, which will become a hugely valuable resource for the Geoscience community that allows more efficient planning of field and laboratory projects and facilitates sharing of samples, which will help build more comprehensive data sets for individual samples. The SESAR catalog will provide links to sample profiles on external systems that hold data about samples, thereby enabling users to easily obtain complete information about samples.
Micro X-ray CT Imaging of Sediments under Confining Pressure
NASA Astrophysics Data System (ADS)
Schindler, M.; Prasad, M.
2016-12-01
We developed a pressure and temperature control system for use inside the micro X-ray CT scanner Xradia 400. We succeeded in building a pressure vessel that can be pressurized to 34.5 MPa (5000 psi) while being transparent to X-rays. The setup can currently be cooled to -5°C and heated to 40°C. We were able to observe grain damage and porosity reduction due to applied confining pressure in clean quartz sand samples and quartz sand and bentonite samples. By comparing micro CT images at atmospheric pressure and 13.8 MPa (2000 psi) confining pressure, we observed compaction of the samples resulting in grain damage and fracturing of sediment grains (Figure 1). When the confining pressure was decreased some grains experienced further fracturing. The grain damage appears irreversible. Further fracturing of grains in pre-compacted sediment was observed upon repeated confining pressure cycling. We are currently working on feed-throughs for fluid lines and electric wiring to use ultrasonic transducers and pressure control in combination. Further we plan to include pore pressure in addition to confining pressure into the system. The pressure control system in combination with ultrasonic transducers will allow us to visually observe pore scale changes in rock samples while simultaneously identifying their influence on ultrasonic velocities. Such pore-scale changes are usually not taken into account by rock physics models and could help to identify why laboratory data diverges from theoretical models. Further, it is possible to compute compressibility from mCT images at different stress states by image correlation
Design of the scanning mode coated glass color difference online detection system
NASA Astrophysics Data System (ADS)
Bi, Weihong; Zhang, Yu; Wang, Dajiang; Zhang, Baojun; Fu, Guangwei
2008-03-01
A design of scanning mode coated glass color difference online detection system was introduced. The system consisted of color difference data acquirement part and orbit control part. The function of the color difference data acquirement part was to acquire glass spectral reflectance and then processed them to get the color difference value. Using fiber for light guiding, the reflected light from surface of glass was transmitted into light division part, and the dispersive light was imaged on linear CCD, and then the output signals from the CCD was sampled pixel by pixel, and the spectral reflectance of coated glass was obtained finally. Then, the acquired spectral reflectance signals was sent to industrial personal computer through USB interface, using standard color space and color difference formula nominated by International Commission on Illumination (CIE) in 1976 to process these signals, and the reflected color parameter and color difference of coated glass was gained in the end. The function of the orbit control part was to move the detection probe by way of transverse scanning mode above the glass strip, and control the measuring start-stop time of the color difference data acquirement part at the same time. The color difference data acquirement part of the system was put on the orbit which is after annealing area in coated glass production line, and the protected fiber probe was placed on slide of the orbit. Using single chip microcomputer to control transmission mechanism of the slide, which made the slide move by way of transverse scanning mode on the glass strip, meanwhile, the color difference data acquirement part of the system was also controlled by the single chip microcomputer, and it made the acquirement part measure color difference data when the probe reached the needed working speed and required place on the glass strip. The scanning mode coated glass color difference online detection system can measure color parameter and color difference of each transverse point on glass strip, it can also measure lengthways color stability on glass strip. Furthermore, the measuring results can be transmitted to coated control room through intranet, so it is very useful to improve producing technique in time. In addition, equipping necessary marking machine, this system can classify glass board automatically based on the measuring result.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bharoto,; Suparno, Nadi; Putra, Edy Giri Rachman
In 2005, the main computer for data acquisition and control system of Small-angle Neutron Scattering (SANS) BATAN Spectrometer (SMARTer) was replaced since it halted to operate the spectrometer. According to this replacement, the new software for data acquisition and control system has been developed in-house. Visual Basic programming language is used in developing the software. In the last two years, many developments have been made both in the hardware and also the software to conduct the experiment is more effective and efficient. Lately, the previous motor controller card (ISA Card) was replaced with the programmable motor controller card (PCI Card)more » for driving one motor of position sensitive detector (PSD), eight motors of four collimators, and six motors of six pinhole discs. This new control system software makes all motors can be moved simultaneously, then it reduces significantly the consuming time of setting up the instrument before running the experiment. Along with that development, the new data acquisition software under MS Windows operating system is also developed to drive a beam stopper in X-Y directions as well as to read the equipment status such as position of the collimators and PSD, to acquire neutron counts on monitor and PSD detectors, and also to manage 12 samples position automatically. A timer object which is set in one second to read the equipment status via serial port of the computer (RS232C), and general purpose interface board (GPIB) for reading the total counts of each pixel of the PSD from histogram memory was used in this new software. The experiment result displayed in real time on the main window, and the data is saved in the special format for further data reduction and analysis. The new software has been implemented and performed for experiment using a preset count or preset time mode for absolute scattering intensity method.« less
NASA Astrophysics Data System (ADS)
Bharoto, Suparno, Nadi; Putra, Edy Giri Rachman
2015-04-01
In 2005, the main computer for data acquisition and control system of Small-angle Neutron Scattering (SANS) BATAN Spectrometer (SMARTer) was replaced since it halted to operate the spectrometer. According to this replacement, the new software for data acquisition and control system has been developed in-house. Visual Basic programming language is used in developing the software. In the last two years, many developments have been made both in the hardware and also the software to conduct the experiment is more effective and efficient. Lately, the previous motor controller card (ISA Card) was replaced with the programmable motor controller card (PCI Card) for driving one motor of position sensitive detector (PSD), eight motors of four collimators, and six motors of six pinhole discs. This new control system software makes all motors can be moved simultaneously, then it reduces significantly the consuming time of setting up the instrument before running the experiment. Along with that development, the new data acquisition software under MS Windows operating system is also developed to drive a beam stopper in X-Y directions as well as to read the equipment status such as position of the collimators and PSD, to acquire neutron counts on monitor and PSD detectors, and also to manage 12 samples position automatically. A timer object which is set in one second to read the equipment status via serial port of the computer (RS232C), and general purpose interface board (GPIB) for reading the total counts of each pixel of the PSD from histogram memory was used in this new software. The experiment result displayed in real time on the main window, and the data is saved in the special format for further data reduction and analysis. The new software has been implemented and performed for experiment using a preset count or preset time mode for absolute scattering intensity method.
Distributed cyberinfrastructure tools for automated data processing of structural monitoring data
NASA Astrophysics Data System (ADS)
Zhang, Yilan; Kurata, Masahiro; Lynch, Jerome P.; van der Linden, Gwendolyn; Sederat, Hassan; Prakash, Atul
2012-04-01
The emergence of cost-effective sensing technologies has now enabled the use of dense arrays of sensors to monitor the behavior and condition of large-scale bridges. The continuous operation of dense networks of sensors presents a number of new challenges including how to manage such massive amounts of data that can be created by the system. This paper reports on the progress of the creation of cyberinfrastructure tools which hierarchically control networks of wireless sensors deployed in a long-span bridge. The internet-enabled cyberinfrastructure is centrally managed by a powerful database which controls the flow of data in the entire monitoring system architecture. A client-server model built upon the database provides both data-provider and system end-users with secured access to various levels of information of a bridge. In the system, information on bridge behavior (e.g., acceleration, strain, displacement) and environmental condition (e.g., wind speed, wind direction, temperature, humidity) are uploaded to the database from sensor networks installed in the bridge. Then, data interrogation services interface with the database via client APIs to autonomously process data. The current research effort focuses on an assessment of the scalability and long-term robustness of the proposed cyberinfrastructure framework that has been implemented along with a permanent wireless monitoring system on the New Carquinez (Alfred Zampa Memorial) Suspension Bridge in Vallejo, CA. Many data interrogation tools are under development using sensor data and bridge metadata (e.g., geometric details, material properties, etc.) Sample data interrogation clients including those for the detection of faulty sensors, automated modal parameter extraction.
40 CFR 63.1032 - Sampling connection systems standards.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 40 Protection of Environment 10 2010-07-01 2010-07-01 false Sampling connection systems standards... (CONTINUED) National Emission Standards for Equipment Leaks-Control Level 2 Standards § 63.1032 Sampling.... Each sampling connection system shall be equipped with a closed-purge, closed-loop, or closed vent...
Ma, Hongjun; Liu, Huajun; Liu, Fang; Zhang, Huahui; Ci, Lu; Shi, Yi; Lei, Lei
2018-01-01
High-Temperature Superconductors (HTS) are potential materials for high-field magnets, low-loss transmission cables, and Superconducting Magnetic Energy Storage (SMES) due to their high upper critical magnetic field (H c2 ) and critical temperature (T c ). The critical current (I c ) of HTS, which is one of the most important parameters for superconductor application, depends strongly on the magnetic fields and temperatures. A new I c measurement system that can carry out accurate I c measurement for HTS short samples with various temperatures (4.2-80 K), magnetic fields (0-14 T), and angles of the magnetic field (0°-90°) has been developed. The I c measurement system mainly consists of a measurement holder, temperature-control system, background magnet, test cryostat, data acquisition system, and DC power supply. The accuracy of temperature control is better than ±0.1 K over the 20-80 K range and ±0.05 K when measured below 20 K. The maximum current is over 1000 A with a measurement uncertainty of 1%. The system had been successfully used for YBa 2 Cu 3 O 7-x (YBCO) tapes I c determination with different temperatures and magnetic fields.
NASA Astrophysics Data System (ADS)
Ma, Hongjun; Liu, Huajun; Liu, Fang; Zhang, Huahui; Ci, Lu; Shi, Yi; Lei, Lei
2018-01-01
High-Temperature Superconductors (HTS) are potential materials for high-field magnets, low-loss transmission cables, and Superconducting Magnetic Energy Storage (SMES) due to their high upper critical magnetic field (Hc2) and critical temperature (Tc). The critical current (Ic) of HTS, which is one of the most important parameters for superconductor application, depends strongly on the magnetic fields and temperatures. A new Ic measurement system that can carry out accurate Ic measurement for HTS short samples with various temperatures (4.2-80 K), magnetic fields (0-14 T), and angles of the magnetic field (0°-90°) has been developed. The Ic measurement system mainly consists of a measurement holder, temperature-control system, background magnet, test cryostat, data acquisition system, and DC power supply. The accuracy of temperature control is better than ±0.1 K over the 20-80 K range and ±0.05 K when measured below 20 K. The maximum current is over 1000 A with a measurement uncertainty of 1%. The system had been successfully used for YBa2Cu3O7-x(YBCO) tapes Ic determination with different temperatures and magnetic fields.
Water quality data for selected wells in the Coastal Plain of New Jersey, 1996-98
Hibbs, Kathleen L.; Stackelberg, Paul E.; Kauffman, Leon J.; Ayers, Mark A.
2001-01-01
Water-quality data were collected during 1996-98 for 217 wells in New Jersey and 3 wells in New York as part of the U. S. Geological Survey's National Water Quality Assessment Program. Samples were collected for five ground-water surveys that were designed to assess water quality in major aquifer systems, with an emphasis on recently recharged (shallow) ground water associated with present and recent human activities. This report (1) summarizes the hydrogeologic framework in the areas of data collection; (2) describes the objectives and procedures for designing each ground-water survey; (3) summarizes the procedures and protocols for data collec-tion, analysis, and quality control; and (4) lists the concentrations of inorganic constituents, volatile organic compounds, pesticides, nutrients, and trace elements present in the ground-water samples.
AU-FREDI - AUTONOMOUS FREQUENCY DOMAIN IDENTIFICATION
NASA Technical Reports Server (NTRS)
Yam, Y.
1994-01-01
The Autonomous Frequency Domain Identification program, AU-FREDI, is a system of methods, algorithms and software that was developed for the identification of structural dynamic parameters and system transfer function characterization for control of large space platforms and flexible spacecraft. It was validated in the CALTECH/Jet Propulsion Laboratory's Large Spacecraft Control Laboratory. Due to the unique characteristics of this laboratory environment, and the environment-specific nature of many of the software's routines, AU-FREDI should be considered to be a collection of routines which can be modified and reassembled to suit system identification and control experiments on large flexible structures. The AU-FREDI software was originally designed to command plant excitation and handle subsequent input/output data transfer, and to conduct system identification based on the I/O data. Key features of the AU-FREDI methodology are as follows: 1. AU-FREDI has on-line digital filter design to support on-orbit optimal input design and data composition. 2. Data composition of experimental data in overlapping frequency bands overcomes finite actuator power constraints. 3. Recursive least squares sine-dwell estimation accurately handles digitized sinusoids and low frequency modes. 4. The system also includes automated estimation of model order using a product moment matrix. 5. A sample-data transfer function parametrization supports digital control design. 6. Minimum variance estimation is assured with a curve fitting algorithm with iterative reweighting. 7. Robust root solvers accurately factorize high order polynomials to determine frequency and damping estimates. 8. Output error characterization of model additive uncertainty supports robustness analysis. The research objectives associated with AU-FREDI were particularly useful in focusing the identification methodology for realistic on-orbit testing conditions. Rather than estimating the entire structure, as is typically done in ground structural testing, AU-FREDI identifies only the key transfer function parameters and uncertainty bounds that are necessary for on-line design and tuning of robust controllers. AU-FREDI's system identification algorithms are independent of the JPL-LSCL environment, and can easily be extracted and modified for use with input/output data files. The basic approach of AU-FREDI's system identification algorithms is to non-parametrically identify the sampled data in the frequency domain using either stochastic or sine-dwell input, and then to obtain a parametric model of the transfer function by curve-fitting techniques. A cross-spectral analysis of the output error is used to determine the additive uncertainty in the estimated transfer function. The nominal transfer function estimate and the estimate of the associated additive uncertainty can be used for robust control analysis and design. AU-FREDI's I/O data transfer routines are tailored to the environment of the CALTECH/ JPL-LSCL which included a special operating system to interface with the testbed. Input commands for a particular experiment (wideband, narrowband, or sine-dwell) were computed on-line and then issued to respective actuators by the operating system. The operating system also took measurements through displacement sensors and passed them back to the software for storage and off-line processing. In order to make use of AU-FREDI's I/O data transfer routines, a user would need to provide an operating system capable of overseeing such functions between the software and the experimental setup at hand. The program documentation contains information designed to support users in either providing such an operating system or modifying the system identification algorithms for use with input/output data files. It provides a history of the theoretical, algorithmic and software development efforts including operating system requirements and listings of some of the various special purpose subroutines which were developed and optimized for Lahey FORTRAN compilers on IBM PC-AT computers before the subroutines were integrated into the system software. Potential purchasers are encouraged to purchase and review the documentation before purchasing the AU-FREDI software. AU-FREDI is distributed in DEC VAX BACKUP format on a 1600 BPI 9-track magnetic tape (standard media) or a TK50 tape cartridge. AU-FREDI was developed in 1989 and is a copyrighted work with all copyright vested in NASA.
NASA Astrophysics Data System (ADS)
Piasecki, M.; Ji, P.
2014-12-01
Geoscience data comes in many flavors that are determined by type of data such as continous on a grid or mesh or discrete colelcted at point either as one time samples or a stream of data coming of sensors, but coudl also encompass digital files of any time type such text files, WORD or EXCEL documents, or audio and video files. We present a storage facility that is comprsed of 6 nodes each of speciaized to host a certain data type: grid based data (netCDF on a THREDDS server), GIS data (shapefiles using GeoServer), point time series data (CUAHSI ODM), sample data (EDBS), and any digital data (RAMADAA) plus a server fro Remote sensing data and its products. While there is overlap in data type storage capabilities (rasters can go into several of these nodes) we prefer to use dedicated storage facilities that are a) freeware, and b) have a good degree of maturity, and c) have shown their utility for stroing a cetain type. In addition it allows to place these commonly used software stacks and storage solutiosn side-by-side to develop interoprability strategies. We have used a DRUPAL based system to handle user regoistration and authentication, and also use the system for data submission and data search. In support for tis system we developed an extensive controlled vocabulary system that is an amalgamation of various CVs used in the geosciecne community in order to achieve as high a degree of recognition, such the CF conventions, CUAHSI Cvs, , NASA (GCMD), EPA and USGS taxonomies, GEMET, in addition to ontological representations such as SWEET.
Development and testing of a portable wind sensitive directional air sampler
NASA Technical Reports Server (NTRS)
Deyo, J.; Toma, J.; King, R. B.
1975-01-01
A portable wind sensitive directional air sampler was developed as part of an air pollution source identification system. The system is designed to identify sources of air pollution based on the directional collection of field air samples and their analysis for TSP and trace element characteristics. Sources can be identified by analyzing the data on the basis of pattern recognition concepts. The unit, designated Air Scout, receives wind direction signals from an associated wind vane. Air samples are collected on filter slides using a standard high volume air sampler drawing air through a porting arrangement which tracks the wind direction and permits collection of discrete samples. A preset timer controls the length of time each filter is in the sampling position. At the conclusion of the sampling period a new filter is automatically moved into sampling position displacing the previous filter to a storage compartment. Thus the Air Scout may be set up at a field location, loaded with up to 12 filter slides, and left to acquire air samples automatically, according to the wind, at any timer interval desired from 1 to 30 hours.
Application of allflex conservation buffer in illumina genotyping.
de Groot, M; Ras, T; van Haeringen, W A
2016-12-01
This experiment was designed to study if liquid conservation buffer used in the novel Tissue Sampling Technology (TST) from Allflex can be used for Illumina BeadChip genotyping. Ear punches were collected from 6 bovine samples, using both the Tissue Sampling Unit (TSU) as well as the Total Tagger Universal (TTU) collection system. The stability of the liquid conservation buffer was tested by genotyping samples on Illumina BeadChips, incubated at 0, 3, 15, 24, 48, 72, 168, 336, 720 h after sample collection. Additionally, a replenishment study was designed to test how often the liquid conservation buffer could be completely replenished before a significant call rate drop could be observed. Results from the stability study showed an average call rate of 0.993 for samples collected with the TSU system and 0.953 for samples collected with the TTU system, both exceeding the inclusion threshold call rate of 0.85. As an additional control, the identity of the individual animals was confirmed using the International Society of Animal Genetics (ISAG) recommended SNP panel. The replenishment study revealed a slight drop in the sample call rate after replenishing the conservation buffer for the fourth time for the TSU as well as the TTU samples. In routine analysis, this application allows for multiple experiments to be performed on the liquid conservation buffer, while maintaining the tissue samples for future use. The data collected in this study shows that the liquid conservation buffer used in the TST system can be used for Illumina BeadChip genotyping applications.
Integrated control-system design via generalized LQG (GLQG) theory
NASA Technical Reports Server (NTRS)
Bernstein, Dennis S.; Hyland, David C.; Richter, Stephen; Haddad, Wassim M.
1989-01-01
Thirty years of control systems research has produced an enormous body of theoretical results in feedback synthesis. Yet such results see relatively little practical application, and there remains an unsettling gap between classical single-loop techniques (Nyquist, Bode, root locus, pole placement) and modern multivariable approaches (LQG and H infinity theory). Large scale, complex systems, such as high performance aircraft and flexible space structures, now demand efficient, reliable design of multivariable feedback controllers which optimally tradeoff performance against modeling accuracy, bandwidth, sensor noise, actuator power, and control law complexity. A methodology is described which encompasses numerous practical design constraints within a single unified formulation. The approach, which is based upon coupled systems or modified Riccati and Lyapunov equations, encompasses time-domain linear-quadratic-Gaussian theory and frequency-domain H theory, as well as classical objectives such as gain and phase margin via the Nyquist circle criterion. In addition, this approach encompasses the optimal projection approach to reduced-order controller design. The current status of the overall theory will be reviewed including both continuous-time and discrete-time (sampled-data) formulations.
System and method for non-destructive evaluation of surface characteristics of a magnetic material
Jiles, David C.; Sipahi, Levent B.
1994-05-17
A system and a related method for non-destructive evaluation of the surface characteristics of a magnetic material. The sample is excited by an alternating magnetic field. The field frequency, amplitude and offset are controlled according to a predetermined protocol. The Barkhausen response of the sample is detected for the various fields and offsets and is analyzed. The system produces information relating to the frequency content, the amplitude content, the average or RMS energy content, as well as count rate information, for each of the Barkhausen responses at each of the excitation levels applied during the protocol. That information provides a contiguous body of data, heretofore unavailable, which can be analyzed to deduce information about the surface characteristics of the material at various depths below the surface.
NASA Technical Reports Server (NTRS)
Hoge, F. E.; Kincaid, J. S.
1980-01-01
A coaxial dual-channel laser system has been developed for the measurement of extinction coefficients of highly absorbing liquids. An empty wedge-shaped sample cell is first translated laterally through a He-Ne laser beam to measure the differential thickness using interference fringes in reflection. The wedge cell is carefully filled with the oil sample and translated through the coaxially positioned dye laser beam for the differential attenuation or extinction measurement. Optional use of the instrumentation as a single-channel extinction measurement system and also as a refractometer is detailed. The system and calibration techniques were applied to the measurement of two crude oils whose extinction values were required to complete the analysis of airborne laser data gathered over four controlled spills.
A Novel Hyperspectral Microscopic Imaging System for Evaluating Fresh Degree of Pork.
Xu, Yi; Chen, Quansheng; Liu, Yan; Sun, Xin; Huang, Qiping; Ouyang, Qin; Zhao, Jiewen
2018-04-01
This study proposed a rapid microscopic examination method for pork freshness evaluation by using the self-assembled hyperspectral microscopic imaging (HMI) system with the help of feature extraction algorithm and pattern recognition methods. Pork samples were stored for different days ranging from 0 to 5 days and the freshness of samples was divided into three levels which were determined by total volatile basic nitrogen (TVB-N) content. Meanwhile, hyperspectral microscopic images of samples were acquired by HMI system and processed by the following steps for the further analysis. Firstly, characteristic hyperspectral microscopic images were extracted by using principal component analysis (PCA) and then texture features were selected based on the gray level co-occurrence matrix (GLCM). Next, features data were reduced dimensionality by fisher discriminant analysis (FDA) for further building classification model. Finally, compared with linear discriminant analysis (LDA) model and support vector machine (SVM) model, good back propagation artificial neural network (BP-ANN) model obtained the best freshness classification with a 100 % accuracy rating based on the extracted data. The results confirm that the fabricated HMI system combined with multivariate algorithms has ability to evaluate the fresh degree of pork accurately in the microscopic level, which plays an important role in animal food quality control.
A Novel Hyperspectral Microscopic Imaging System for Evaluating Fresh Degree of Pork
Xu, Yi; Chen, Quansheng; Liu, Yan; Sun, Xin; Huang, Qiping; Ouyang, Qin; Zhao, Jiewen
2018-01-01
Abstract This study proposed a rapid microscopic examination method for pork freshness evaluation by using the self-assembled hyperspectral microscopic imaging (HMI) system with the help of feature extraction algorithm and pattern recognition methods. Pork samples were stored for different days ranging from 0 to 5 days and the freshness of samples was divided into three levels which were determined by total volatile basic nitrogen (TVB-N) content. Meanwhile, hyperspectral microscopic images of samples were acquired by HMI system and processed by the following steps for the further analysis. Firstly, characteristic hyperspectral microscopic images were extracted by using principal component analysis (PCA) and then texture features were selected based on the gray level co-occurrence matrix (GLCM). Next, features data were reduced dimensionality by fisher discriminant analysis (FDA) for further building classification model. Finally, compared with linear discriminant analysis (LDA) model and support vector machine (SVM) model, good back propagation artificial neural network (BP-ANN) model obtained the best freshness classification with a 100 % accuracy rating based on the extracted data. The results confirm that the fabricated HMI system combined with multivariate algorithms has ability to evaluate the fresh degree of pork accurately in the microscopic level, which plays an important role in animal food quality control. PMID:29805285
Winer, Rachel L; Tiro, Jasmin A; Miglioretti, Diana L; Thayer, Chris; Beatty, Tara; Lin, John; Gao, Hongyuan; Kimbel, Kilian; Buist, Diana S M
2018-01-01
Women who delay or do not attend Papanicolaou (Pap) screening are at increased risk for cervical cancer. Trials in countries with organized screening programs have demonstrated that mailing high-risk (hr) human papillomavirus (HPV) self-sampling kits to under-screened women increases participation, but U.S. data are lacking. HOME is a pragmatic randomized controlled trial set within a U.S. integrated healthcare delivery system to compare two programmatic approaches for increasing cervical cancer screening uptake and effectiveness in under-screened women (≥3.4years since last Pap) aged 30-64years: 1) usual care (annual patient reminders and ad hoc outreach by clinics) and 2) usual care plus mailed hrHPV self-screening kits. Over 2.5years, eligible women were identified through electronic medical record (EMR) data and randomized 1:1 to the intervention or control arm. Women in the intervention arm were mailed kits with pre-paid envelopes to return samples to the central clinical laboratory for hrHPV testing. Results were documented in the EMR to notify women's primary care providers of appropriate follow-up. Primary outcomes are detection and treatment of cervical neoplasia. Secondary outcomes are cervical cancer screening uptake, abnormal screening results, and women's experiences and attitudes towards hrHPV self-sampling and follow-up of hrHPV-positive results (measured through surveys and interviews). The trial was designed to evaluate whether a programmatic strategy incorporating hrHPV self-sampling is effective in promoting adherence to the complete screening process (including follow-up of abnormal screening results and treatment). The objective of this report is to describe the rationale and design of this pragmatic trial. Copyright © 2017 Elsevier Inc. All rights reserved.
Fully Automated Data Collection Using PAM and the Development of PAM/SPACE Reversible Cassettes
NASA Astrophysics Data System (ADS)
Hiraki, Masahiko; Watanabe, Shokei; Chavas, Leonard M. G.; Yamada, Yusuke; Matsugaki, Naohiro; Igarashi, Noriyuki; Wakatsuki, Soichi; Fujihashi, Masahiro; Miki, Kunio; Baba, Seiki; Ueno, Go; Yamamoto, Masaki; Suzuki, Mamoru; Nakagawa, Atsushi; Watanabe, Nobuhisa; Tanaka, Isao
2010-06-01
To remotely control and automatically collect data in high-throughput X-ray data collection experiments, the Structural Biology Research Center at the Photon Factory (PF) developed and installed sample exchange robots PAM (PF Automated Mounting system) at PF macromolecular crystallography beamlines; BL-5A, BL-17A, AR-NW12A and AR-NE3A. We developed and installed software that manages the flow of the automated X-ray experiments; sample exchanges, loop-centering and X-ray diffraction data collection. The fully automated data collection function has been available since February 2009. To identify sample cassettes, PAM employs a two-dimensional bar code reader. New beamlines, BL-1A at the Photon Factory and BL32XU at SPring-8, are currently under construction as part of Targeted Proteins Research Program (TPRP) by the Ministry of Education, Culture, Sports, Science and Technology of Japan. However, different robots, PAM and SPACE (SPring-8 Precise Automatic Cryo-sample Exchanger), will be installed at BL-1A and BL32XU, respectively. For the convenience of the users of both facilities, pins and cassettes for PAM and SPACE are developed as part of the TPRP.
NASA Technical Reports Server (NTRS)
Aucoin, P. J.; Stewart, J.; Mckay, M. F. (Principal Investigator)
1980-01-01
This document presents instructions for analysts who use the EOD-LARSYS as programmed on the Purdue University IBM 370/148 (recently replaced by the IBM 3031) computer. It presents sample applications, control cards, and error messages for all processors in the system and gives detailed descriptions of the mathematical procedures and information needed to execute the system and obtain the desired output. EOD-LARSYS is the JSC version of an integrated batch system for analysis of multispectral scanner imagery data. The data included is designed for use with the as built documentation (volume 3) and the program listings (volume 4). The system is operational from remote terminals at Johnson Space Center under the virtual machine/conversational monitor system environment.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tarisien, M.; Plaisir, C.; Gobet, F.
2011-02-15
We present a stand-alone system to characterize the high-energy particles emitted in the interaction of ultrahigh intensity laser pulses with matter. According to the laser and target characteristics, electrons or protons are produced with energies higher than a few mega electron volts. Selected material samples can, therefore, be activated via nuclear reactions. A multidetector, named NATALIE, has been developed to count the {beta}{sup +} activity of these irradiated samples. The coincidence technique used, designed in an integrated system, results in very low background in the data, which is required for low activity measurements. It, therefore, allows a good precision onmore » the nuclear activation yields of the produced radionuclides. The system allows high counting rates and online correction of the dead time. It also provides, online, a quick control of the experiment. Geant4 simulations are used at different steps of the data analysis to deduce, from the measured activities, the energy and angular distributions of the laser-induced particle beams. Two applications are presented to illustrate the characterization of electrons and protons.« less
A compact electroencephalogram recording device with integrated audio stimulation system.
Paukkunen, Antti K O; Kurttio, Anttu A; Leminen, Miika M; Sepponen, Raimo E
2010-06-01
A compact (96 x 128 x 32 mm(3), 374 g), battery-powered, eight-channel electroencephalogram recording device with an integrated audio stimulation system and a wireless interface is presented. The recording device is capable of producing high-quality data, while the operating time is also reasonable for evoked potential studies. The effective measurement resolution is about 4 nV at 200 Hz sample rate, typical noise level is below 0.7 microV(rms) at 0.16-70 Hz, and the estimated operating time is 1.5 h. An embedded audio decoder circuit reads and plays wave sound files stored on a memory card. The activities are controlled by an 8 bit main control unit which allows accurate timing of the stimuli. The interstimulus interval jitter measured is less than 1 ms. Wireless communication is made through bluetooth and the data recorded are transmitted to an external personal computer (PC) interface in real time. The PC interface is implemented with LABVIEW and in addition to data acquisition it also allows online signal processing, data storage, and control of measurement activities such as contact impedance measurement, for example. The practical application of the device is demonstrated in mismatch negativity experiment with three test subjects.
A compact electroencephalogram recording device with integrated audio stimulation system
NASA Astrophysics Data System (ADS)
Paukkunen, Antti K. O.; Kurttio, Anttu A.; Leminen, Miika M.; Sepponen, Raimo E.
2010-06-01
A compact (96×128×32 mm3, 374 g), battery-powered, eight-channel electroencephalogram recording device with an integrated audio stimulation system and a wireless interface is presented. The recording device is capable of producing high-quality data, while the operating time is also reasonable for evoked potential studies. The effective measurement resolution is about 4 nV at 200 Hz sample rate, typical noise level is below 0.7 μVrms at 0.16-70 Hz, and the estimated operating time is 1.5 h. An embedded audio decoder circuit reads and plays wave sound files stored on a memory card. The activities are controlled by an 8 bit main control unit which allows accurate timing of the stimuli. The interstimulus interval jitter measured is less than 1 ms. Wireless communication is made through bluetooth and the data recorded are transmitted to an external personal computer (PC) interface in real time. The PC interface is implemented with LABVIEW® and in addition to data acquisition it also allows online signal processing, data storage, and control of measurement activities such as contact impedance measurement, for example. The practical application of the device is demonstrated in mismatch negativity experiment with three test subjects.
Niesen, Shelley L.; Christensen, Eric D.
2015-01-01
Water-quality, hydrological, and ecological data collected from June 2005 through September 2013 from the Little Blue River and smaller streams within the City of Independence, Missouri, are presented in this report. These data were collected as a part of an ongoing cooperative study between the U.S. Geological Survey and the City of Independence Water Pollution Control Department to characterize the water quality and ecological condition of Independence streams. The quantities, sources of selected constituents, and processes affecting water quality and aquatic life were evaluated to determine the resulting ecological condition of streams within Independence. Data collected for this study fulfill the municipal separate sewer system permit requirements for the City of Independence and can be used to provide a baseline with which city managers can determine the effectiveness of current (2014) and future best management practices within Independence. Continuous streamflow and water-quality data, collected during base flow and stormflow, included physical and chemical properties, inorganic constituents, common organic micro-constituents, pesticides in streambed sediment and surface water, fecal indicator bacteria and microbial source tracking data, and suspended sediment. Dissolved oxygen, pH, specific conductance, water temperature, and turbidity data were measured continuously at seven sites within Independence. Base-flow and stormflow samples were collected at eight gaged and two ungaged sites. Fecal sources samples were collected for reference for microbial source tracking, and sewage influent samples were collected as additional source samples. Dry-weather screening was done on 11 basins within Independence to identify potential contaminant sources to the streams. Benthic macroinvertebrate community surveys and habitat assessments were done on 10 stream sites and 2 comparison sites outside the city. Sampling and laboratory procedures and quality-assurance and quality-control methods used in data collection for this study are described in this report.
Access control for electronic patient records.
Glagola, M J
1998-01-01
The transition from hardcopy records to electronic records is in the forefront for healthcare today. For healthcare facilities, a major issue is determining who can access patients' medical information and how access to this information can be controlled. There are three components to access control: identification, authentication and authorization. Checking proof of identity is a means of authenticating someone--through a driver's license, passport or their fingerprints. Similar processes are needed in a computer environment, through the use of passwords, one-time passwords or smartcards, encryption and kerberos, and call-back procedures. New in the area of access control are biometric devices, which are hardware/software combinations that digitize a physical characteristic and compare the sample with previously stored samples. Fingerprints, voiceprints and facial features are examples. Their cost is currently prohibitive, but in time, they may become more common. Digital certificates and certification authorities are other means used to authenticate identify. When a system challenges a user's identity at log on, the user provides a certification that tells the system to go to the issuing certification authority and find proof the user's claim is valid. Low-level certifications offer little value for sensitive data, but high-level certification is now being introduced. It requires more specific, detailed information on the applicant. Authorization, the final component of access control, establishes what a specific user can and cannot access. To have effective access control, transaction logging and system monitoring are needed to ensure the various techniques are being used and performing properly.
System automatically supplies precise analytical samples of high-pressure gases
NASA Technical Reports Server (NTRS)
Langdon, W. M.
1967-01-01
High-pressure-reducing and flow-stabilization system delivers analytical gas samples from a gas supply. The system employs parallel capillary restrictors for pressure reduction and downstream throttling valves for flow control. It is used in conjunction with a sampling valve and minimizes alterations of the sampled gas.
Automated position control of a surface array relative to a liquid microjunction surface sampler
Van Berkel, Gary J.; Kertesz, Vilmos; Ford, Michael James
2007-11-13
A system and method utilizes an image analysis approach for controlling the probe-to-surface distance of a liquid junction-based surface sampling system for use with mass spectrometric detection. Such an approach enables a hands-free formation of the liquid microjunction used to sample solution composition from the surface and for re-optimization, as necessary, of the microjunction thickness during a surface scan to achieve a fully automated surface sampling system.
Jurek, Anne M; Maldonado, George; Greenland, Sander
2013-03-01
Special care must be taken when adjusting for outcome misclassification in case-control data. Basic adjustment formulas using either sensitivity and specificity or predictive values (as with external validation data) do not account for the fact that controls are sampled from a much larger pool of potential controls. A parallel problem arises in surveys and cohort studies in which participation or loss is outcome related. We review this problem and provide simple methods to adjust for outcome misclassification in case-control studies, and illustrate the methods in a case-control birth certificate study of cleft lip/palate and maternal cigarette smoking during pregnancy. Adjustment formulas for outcome misclassification that ignore case-control sampling can yield severely biased results. In the data we examined, the magnitude of error caused by not accounting for sampling is small when population sensitivity and specificity are high, but increases as (1) population sensitivity decreases, (2) population specificity decreases, and (3) the magnitude of the differentiality increases. Failing to account for case-control sampling can result in an odds ratio adjusted for outcome misclassification that is either too high or too low. One needs to account for outcome-related selection (such as case-control sampling) when adjusting for outcome misclassification using external information. Copyright © 2013 Elsevier Inc. All rights reserved.
Design and Performance of a Spectrometer for Deployment on MISSE 7
NASA Technical Reports Server (NTRS)
Pippin, Gary; Beymer, Jim; Robb, Andrew; Longino, James; Perry, George; Stewart, Alan; Finkenor, Miria
2009-01-01
A spectrometer for reflectance and transmission measurements of samples exposed to the space environment has been developed for deployment on the Materials on the International Space Station Experiment (MISSE) 7. The instrument incorporates a miniature commercial fiber optic coupled spectrometer with a computer control system for detector operation, sample motion and illumination. A set of three spectrometers were recently integrated on the MISSE7 platform with launch and deployment on the International Space Station scheduled for summer of this year. The instrument is one of many active experiments on the platform. The performance of the instrument prior to launch will be discussed. Data from samples measured in the laboratory will be compared to those from the instrument prior to launch. These comparisons will illustrate the capabilities of the current design. The space environment challenges many materials. When in operation on the MISSE 7 platform, the new spectrometer will provide real time data on the how the space environment affects the optical properties of thermal control paints and optical coatings. Data obtained from comparison of pre and post flight measurements on hundreds of samples exposed on previous MISSE platforms have been reported at these meetings. With the new spectrometer and the ability to correlate measured changes with time on orbit and the occurrence of both natural events and human activities, a better understanding of the processes responsible for degradation of materials in space will be possible.
NASA Technical Reports Server (NTRS)
Mukhopadhyay, V.
1988-01-01
A generic procedure for the parameter optimization of a digital control law for a large-order flexible flight vehicle or large space structure modeled as a sampled data system is presented. A linear quadratic Guassian type cost function was minimized, while satisfying a set of constraints on the steady-state rms values of selected design responses, using a constrained optimization technique to meet multiple design requirements. Analytical expressions for the gradients of the cost function and the design constraints on mean square responses with respect to the control law design variables are presented.
NASA Technical Reports Server (NTRS)
Rosen, I. G.; Wang, C.
1990-01-01
The convergence of solutions to the discrete or sampled time linear quadratic regulator problem and associated Riccati equation for infinite dimensional systems to the solutions to the corresponding continuous time problem and equation, as the length of the sampling interval (the sampling rate) tends toward zero (infinity) is established. Both the finite and infinite time horizon problems are studied. In the finite time horizon case, strong continuity of the operators which define the control system and performance index together with a stability and consistency condition on the sampling scheme are required. For the infinite time horizon problem, in addition, the sampled systems must be stabilizable and detectable, uniformly with respect to the sampling rate. Classes of systems for which this condition can be verified are discussed. Results of numerical studies involving the control of a heat/diffusion equation, a hereditary of delay system, and a flexible beam are presented and discussed.
NASA Technical Reports Server (NTRS)
Rosen, I. G.; Wang, C.
1992-01-01
The convergence of solutions to the discrete- or sampled-time linear quadratic regulator problem and associated Riccati equation for infinite-dimensional systems to the solutions to the corresponding continuous time problem and equation, as the length of the sampling interval (the sampling rate) tends toward zero(infinity) is established. Both the finite-and infinite-time horizon problems are studied. In the finite-time horizon case, strong continuity of the operators that define the control system and performance index, together with a stability and consistency condition on the sampling scheme are required. For the infinite-time horizon problem, in addition, the sampled systems must be stabilizable and detectable, uniformly with respect to the sampling rate. Classes of systems for which this condition can be verified are discussed. Results of numerical studies involving the control of a heat/diffusion equation, a hereditary or delay system, and a flexible beam are presented and discussed.
W-026, transuranic waste restricted waste management (TRU RWM) glovebox operational test report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Leist, K.J.
1998-02-18
The TRU Waste/Restricted Waste Management (LLW/PWNP) Glovebox 401 is designed to accept and process waste from the Transuranic Process Glovebox 302. Waste is transferred to the glovebox via the Drath and Schraeder Bagless Transfer Port (DO-07401) on a transfer stand. The stand is removed with a hoist and the operator inspects the waste (with the aid of the Sampling and Treatment Director) to determine a course of action for each item. The waste is separated into compliant and non compliant. One Trip Port DO-07402A is designated as ``Compliant``and One Trip Port DO-07402B is designated as ``Non Compliant``. As the processingmore » (inspection, bar coding, sampling and treatment) of the transferred items takes place, residue is placed in the appropriate One Trip port. The status of the waste items is tracked by the Data Management System (DMS) via the Plant Control System (PCS) barcode interface. As an item is moved for sampling or storage or it`s state altered by treatment, the Operator will track an items location using a portable barcode reader and entry any required data on the DMS console. The Operational Test Procedure (OTP) will perform evolutions (described here) using the Plant Operating Procedures (POP) in order to verify that they are sufficient and accurate for controlled glovebox operation.« less
Evaluation and Comparison of Methods for Measuring Ozone ...
Ambient evaluations of the various ozone and NO2 methods were conducted during field intensive studies as part of the NASA DISCOVER-AQ project conducted during July 2011 near Baltimore, MD; January – February 2013 in the San Juaquin valley, CA; September 2013 in Houston, TX; and July – August 2014 near Denver, CO. During field intensive studies, instruments were calibrated according to manufacturers’ operation manuals and in accordance with FRM requirements listed in 40 CFR 50. During the ambient evaluation campaigns, nightly automated zero and span checks were performed to monitor the validity of the calibration and control for drifts or variations in the span and/or zero response. Both the calibration gas concentrations and the nightly zero and span gas concentrations were delivered using a dynamic dilution calibration system (T700U/T701H, Teledyne API). The analyzers were housed within a temperature-controlled shelter during the sampling campaigns. A glass inlet with sampling height located approximately 5 m above ground level and a subsequent sampling manifold were shared by all instruments. Data generated by all analyzers were collected and logged using a field deployable data acquisition system (Envidas Ultimate). A summary of instruments used during DISCOVER-AQ deployment are listed in Table 1. Figure 1 shows a typical DISCOVER-AQ site (Houston 2013) where EPA (and others) instrumentation was deployed. Under the Clean Air Act, the U.S. EPA has estab
Lu, Binglong; Jiang, Haijun; Hu, Cheng; Abdurahman, Abdujelil
2018-05-04
The exponential synchronization of hybrid coupled reaction-diffusion neural networks with time delays is discussed in this article. At first, a generalized intermittent control with spacial sampled-data is introduced, which is intermittent in time and data sampling in space. This type of control strategy not only can unify the traditional periodic intermittent control and the aperiodic case, but also can lower the update rate of the controller in both temporal and spatial domains. Next, based on the designed control protocol and the Lyapunov-Krasovskii functional approach, some novel and readily verified criteria are established to guarantee the exponential synchronization of the considered networks. These criteria depend on the diffusion coefficients, coupled strengths, time delays as well as control parameters. Finally, the effectiveness of the proposed control strategy is shown by a numerical example. Copyright © 2018 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Valentic, T. A.
2012-12-01
The Data Transport Network is designed for the delivery of data from scientific instruments located at remote field sites with limited or unreliable communications. Originally deployed at the Sondrestrom Research Facility in Greenland over a decade ago, the system supports the real-time collection and processing of data from large instruments such as incoherent scatter radars and lidars. In recent years, the Data Transport Network has been adapted to small, low-power embedded systems controlling remote instrumentation platforms deployed throughout the Arctic. These projects include multiple buoys from the O-Buoy, IceLander and IceGoat programs, renewable energy monitoring at the Imnavait Creek and Ivotuk field sites in Alaska and remote weather observation stations in Alaska and Greenland. This presentation will discuss the common communications controller developed for these projects. Although varied in their application, each of these systems share a number of common features. Multiple instruments are attached, each of which needs to be power controlled, data sampled and files transmitted offsite. In addition, the power usage of the overall system must be minimized to handle the limited energy available from sources such as solar, wind and fuel cells. The communications links are satellite based. The buoys and weather stations utilize Iridium, necessitating the need to handle the common drop outs and high-latency, low-bandwidth nature of the link. The communications controller is an off-the-shelf, low-power, single board computer running a customized version of the Linux operating system. The Data Transport Network provides a Python-based software framework for writing individual data collection programs and supplies a number of common services for configuration, scheduling, logging, data transmission and resource management. Adding a new instrument involves writing only the necessary code for interfacing to the hardware. Individual programs communicate with the system services using XML-RPC. The scheduling algorithms have access the current position and power levels, allowing for instruments such as cameras to only be run during daylight hours or when sufficient power is available. The resource manager monitors the use of common devices such as the USB bus or Ethernet ports, and can power them down when they are not being used. This management lets us drop the power consumption from an average of 1W to 250mW.
Kawada, Y; Yamada, T; Unno, Y; Yunoki, A; Sato, Y; Hino, Y
2012-09-01
A simple but versatile data acquisition system for software coincidence experiments is described, in which any time stamping and live time controller are not provided. Signals from β- and γ-channels are fed to separately two fast ADCs (16 bits, 25 MHz clock maximum) via variable delay circuits and pulse-height stretchers, and also to pulse-height discriminators. The discriminating level was set to just above the electronic noise. Two ADCs were controlled with a common clock signal, and triggered simultaneously by the logic OR pulses from both discriminators. Paired digital signals for each sampling were sent to buffer memories connected to main PC with a FIFO (First-In, First-Out) pipe via USB. After data acquisition in list mode, various processing including pulse-height analyses was performed using MS-Excel (version 2007 and later). The usefulness of this system was demonstrated for 4πβ(PS)-4πγ coincidence measurements of (60)Co, (134)Cs and (152)Eu. Possibilities of other extended applications will be touched upon. Copyright © 2012 Elsevier Ltd. All rights reserved.
Menheer, M.A.; Brigham, M.E.
1997-01-01
Quality-control data demonstrated that most constituents measured for this study yielded reproducible data, with low to undetectable contamination from the sampling and analytical procedures. Several constituents were occasionally or frequently detected in blank samples at levels similar to low-concentration ground-water-quality samples. For example, iron was detected in 75 percent of the blank samples, with a maximum concentration of 27 [ig/L, indicating that iron contamination may interfere with its determination at low levels in ground waters. Copper, aluminum, and dissolved organic carbon concentrations in blank samples overlap those determined in ground-waterquality samples, thereby precluding quantitative reporting of those constituents. Most pesticide data are reproducible, with minimal bias. Some pesticides had low but consistent recoveries; these data may be useful if spike and surrogate data are carefully considered. Data for some pesticides measured in this study should not be quantitatively reported or used, because they may underestimate the concentrations of those pesticides in ground waters.
Nonlinear Complexity Analysis of Brain fMRI Signals in Schizophrenia
Sokunbi, Moses O.; Gradin, Victoria B.; Waiter, Gordon D.; Cameron, George G.; Ahearn, Trevor S.; Murray, Alison D.; Steele, Douglas J.; Staff, Roger T.
2014-01-01
We investigated the differences in brain fMRI signal complexity in patients with schizophrenia while performing the Cyberball social exclusion task, using measures of Sample entropy and Hurst exponent (H). 13 patients meeting diagnostic and Statistical Manual of Mental Disorders, 4th Edition (DSM IV) criteria for schizophrenia and 16 healthy controls underwent fMRI scanning at 1.5 T. The fMRI data of both groups of participants were pre-processed, the entropy characterized and the Hurst exponent extracted. Whole brain entropy and H maps of the groups were generated and analysed. The results after adjusting for age and sex differences together show that patients with schizophrenia exhibited higher complexity than healthy controls, at mean whole brain and regional levels. Also, both Sample entropy and Hurst exponent agree that patients with schizophrenia have more complex fMRI signals than healthy controls. These results suggest that schizophrenia is associated with more complex signal patterns when compared to healthy controls, supporting the increase in complexity hypothesis, where system complexity increases with age or disease, and also consistent with the notion that schizophrenia is characterised by a dysregulation of the nonlinear dynamics of underlying neuronal systems. PMID:24824731
Data-quality measures for stakeholder-implemented watershed-monitoring programs
Greve, Adrienne I.
2002-01-01
Community-based watershed groups, many of which collect environmental data, have steadily increased in number over the last decade. The data generated by these programs are often underutilized due to uncertainty in the quality of data produced. The incorporation of data-quality measures into stakeholder monitoring programs lends statistical validity to data. Data-quality measures are divided into three steps: quality assurance, quality control, and quality assessment. The quality-assurance step attempts to control sources of error that cannot be directly quantified. This step is part of the design phase of a monitoring program and includes clearly defined, quantifiable objectives, sampling sites that meet the objectives, standardized protocols for sample collection, and standardized laboratory methods. Quality control (QC) is the collection of samples to assess the magnitude of error in a data set due to sampling, processing, transport, and analysis. In order to design a QC sampling program, a series of issues needs to be considered: (1) potential sources of error, (2) the type of QC samples, (3) inference space, (4) the number of QC samples, and (5) the distribution of the QC samples. Quality assessment is the process of evaluating quality-assurance measures and analyzing the QC data in order to interpret the environmental data. Quality assessment has two parts: one that is conducted on an ongoing basis as the monitoring program is running, and one that is conducted during the analysis of environmental data. The discussion of the data-quality measures is followed by an example of their application to a monitoring program in the Big Thompson River watershed of northern Colorado.
One GHz digitizer for space based laser altimeter
NASA Technical Reports Server (NTRS)
Staples, Edward J.
1991-01-01
This is the final report for the research and development of the one GHz digitizer for space based laser altimeter. A feasibility model was designed, built, and tested. Only partial testing of essential functions of the digitizer was completed. Hybrid technology was incorporated which allows analog storage (memory) of the digitally sampled data. The actual sampling rate is 62.5 MHz, but executed in 16 parallel channels, to provide an effective sampling rate of one GHz. The average power consumption of the one GHz digitizer is not more than 1.5 Watts. A one GHz oscillator is incorporated for timing purposes. This signal is also made available externally for system timing. A software package was also developed for internal use (controls, commands, etc.) and for data communication with the host computer. The digitizer is equipped with an onboard microprocessor for this purpose.
Fiebig, Lukas; Laux, Ralf; Binder, Rudolf; Ebner, Thomas
2016-10-01
1. Liquid chromatography (LC)-high resolution mass spectrometry (HRMS) techniques proved to be well suited for the identification of predicted and unexpected drug metabolites in complex biological matrices. 2. To efficiently discriminate between drug-related and endogenous matrix compounds, however, sophisticated postacquisition data mining tools, such as control comparison techniques are needed. For preclinical absorption, distribution, metabolism and excretion (ADME) studies that usually lack a placebo-dosed control group, the question arises how high-quality control data can be yielded using only a minimum number of control animals. 3. In the present study, the combination of LC-traveling wave ion mobility separation (TWIMS)-HRMS(E) and multivariate data analysis was used to study the polymer patterns of the frequently used formulation constituents polyethylene glycol 400 and polysorbate 80 in rat plasma and urine after oral and intravenous administration, respectively. 4. Complex peak patterns of both constituents were identified underlining the general importance of a vehicle-dosed control group in ADME studies for control comparison. Furthermore, the detailed analysis of administration route, blood sampling time and gender influences on both vehicle peak pattern as well as endogenous matrix background revealed that high-quality control data is obtained when (i) control animals receive an intravenous dose of the vehicle, (ii) the blood sampling time point is the same for analyte and control sample and (iii) analyte and control samples of the same gender are compared.