NASA Technical Reports Server (NTRS)
White, D. R.
1976-01-01
A high-vacuum complex composed of an atmospheric decontamination system, sample-processing chambers, storage chambers, and a transfer system was built to process and examine lunar material while maintaining quarantine status. Problems identified, equipment modifications, and procedure changes made for Apollo 11 and 12 sample processing are presented. The sample processing experiences indicate that only a few operating personnel are required to process the sample efficiently, safely, and rapidly in the high-vacuum complex. The high-vacuum complex was designed to handle the many contingencies, both quarantine and scientific, associated with handling an unknown entity such as the lunar sample. Lunar sample handling necessitated a complex system that could not respond rapidly to changing scientific requirements as the characteristics of the lunar sample were better defined. Although the complex successfully handled the processing of Apollo 11 and 12 lunar samples, the scientific requirement for vacuum samples was deleted after the Apollo 12 mission just as the vacuum system was reaching its full potential.
Microfluidic-Based Robotic Sampling System for Radioactive Solutions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jack D. Law; Julia L. Tripp; Tara E. Smith
A novel microfluidic based robotic sampling system has been developed for sampling and analysis of liquid solutions in nuclear processes. This system couples the use of a microfluidic sample chip with a robotic system designed to allow remote, automated sampling of process solutions in-cell and facilitates direct coupling of the microfluidic sample chip with analytical instrumentation. This system provides the capability for near real time analysis, reduces analytical waste, and minimizes the potential for personnel exposure associated with traditional sampling methods. A prototype sampling system was designed, built and tested. System testing demonstrated operability of the microfluidic based sample systemmore » and identified system modifications to optimize performance.« less
Non-Contact Conductivity Measurement for Automated Sample Processing Systems
NASA Technical Reports Server (NTRS)
Beegle, Luther W.; Kirby, James P.
2012-01-01
A new method has been developed for monitoring and control of automated sample processing and preparation especially focusing on desalting of samples before analytical analysis (described in more detail in Automated Desalting Apparatus, (NPO-45428), NASA Tech Briefs, Vol. 34, No. 8 (August 2010), page 44). The use of non-contact conductivity probes, one at the inlet and one at the outlet of the solid phase sample preparation media, allows monitoring of the process, and acts as a trigger for the start of the next step in the sequence (see figure). At each step of the muti-step process, the system is flushed with low-conductivity water, which sets the system back to an overall low-conductivity state. This measurement then triggers the next stage of sample processing protocols, and greatly minimizes use of consumables. In the case of amino acid sample preparation for desalting, the conductivity measurement will define three key conditions for the sample preparation process. First, when the system is neutralized (low conductivity, by washing with excess de-ionized water); second, when the system is acidified, by washing with a strong acid (high conductivity); and third, when the system is at a basic condition of high pH (high conductivity). Taken together, this non-contact conductivity measurement for monitoring sample preparation will not only facilitate automation of the sample preparation and processing, but will also act as a way to optimize the operational time and use of consumables
Event Processing and Variable Part of Sample Period Determining in Combined Systems Using GA
NASA Astrophysics Data System (ADS)
Strémy, Maximilián; Závacký, Pavol; Jedlička, Martin
2011-01-01
This article deals with combined dynamic systems and usage of modern techniques in dealing with these systems, focusing particularly on sampling period design, cyclic processing tasks and related processing algorithms in the combined event management systems using genetic algorithms.
Gas sampling system for a mass spectrometer
Taylor, Charles E; Ladner, Edward P
2003-12-30
The present invention relates generally to a gas sampling system, and specifically to a gas sampling system for transporting a hazardous process gas to a remotely located mass spectrometer. The gas sampling system includes a capillary tube having a predetermined capillary length and capillary diameter in communication with the supply of process gas and the mass spectrometer, a flexible tube surrounding and coaxial with the capillary tube intermediate the supply of process gas and the mass spectrometer, a heat transfer tube surrounding and coaxial with the capillary tube, and a heating device in communication the heat transfer tube for substantially preventing condensation of the process gas within the capillary tube.
Norton, D M; McCamey, M; Boor, K J; Wiedmann, M
2000-03-01
The cold-smoked fish industry was used as a model for the development of a system for monitoring Listeria spp. in foods and in the food processing environment. A total of 214 samples including raw fish, fish during the cold-smoking process, finished product, and environmental samples were collected from three processing facilities over two visits to each facility. Samples were screened for Listeria spp. using the BAX for Screening/genus Listeria polymerase chain reaction system (PCR) and by culture. Listeria spp., confirmed by the API Listeria test strip or by a PCR assay targeting the L. monocytogenes hlyA gene, were isolated from a total of 89 (41.6%) samples. Of these, 80 samples also tested positive for Listeria spp. using the BAX system. Specifically, 42 (55.3%) environmental samples (n = 76), 11 (25.6%) raw materials samples (n = 43), 20 (35.1%) samples from fish in various stages of processing (n = 57), and 7 (18.4%) finished product samples (n = 38) tested positive for Listeria spp. using the BAX system. Five (4.0%) of the 125 culture-negative samples yielded BAX system-positive results. Listeria isolates from each of nine culture-positive/BAX system-negative samples yielded a positive reaction when tested in pure culture by the BAX system, suggesting that our false-negative results were likely due to the presence of low Listeria numbers in the initial enrichment as opposed to nonreacting isolates. The employment of alternative enrichment protocols, such as the two-step enrichment recommended by the manufacturer, may increase the sensitivity of the assay.
Integrated Multi-process Microfluidic Systems for Automating Analysis
Yang, Weichun; Woolley, Adam T.
2010-01-01
Microfluidic technologies have been applied extensively in rapid sample analysis. Some current challenges for standard microfluidic systems are relatively high detection limits, and reduced resolving power and peak capacity compared to conventional approaches. The integration of multiple functions and components onto a single platform can overcome these separation and detection limitations of microfluidics. Multiplexed systems can greatly increase peak capacity in multidimensional separations and can increase sample throughput by analyzing many samples simultaneously. On-chip sample preparation, including labeling, preconcentration, cleanup and amplification, can all serve to speed up and automate processes in integrated microfluidic systems. This paper summarizes advances in integrated multi-process microfluidic systems for automated analysis, their benefits and areas for needed improvement. PMID:20514343
Smith, Joseph M.; Wells, Sarah P.; Mather, Martha E.; Muth, Robert M.
2014-01-01
When researchers and managers initiate sampling on a new stream or river system, they do not know how effective each gear type is and whether their sampling effort is adequate. Although the types and amount of gear may be different for other studies, systems, and research questions, the five-step process described here for making sampling decisions and evaluating sampling efficiency can be applied widely to any system to restore, manage, and conserve aquatic ecosystems. It is believed that incorporating this gear-evaluation process into a wide variety of studies and ecosystems will increase rigour within and across aquatic biodiversity studies.
Autoverification process improvement by Six Sigma approach: Clinical chemistry & immunoassay.
Randell, Edward W; Short, Garry; Lee, Natasha; Beresford, Allison; Spencer, Margaret; Kennell, Marina; Moores, Zoë; Parry, David
2018-05-01
This study examines effectiveness of a project to enhance an autoverification (AV) system through application of Six Sigma (DMAIC) process improvement strategies. Similar AV systems set up at three sites underwent examination and modification to produce improved systems while monitoring proportions of samples autoverified, the time required for manual review and verification, sample processing time, and examining characteristics of tests not autoverified. This information was used to identify areas for improvement and monitor the impact of changes. Use of reference range based criteria had the greatest impact on the proportion of tests autoverified. To improve AV process, reference range based criteria was replaced with extreme value limits based on a 99.5% test result interval, delta check criteria were broadened, and new specimen consistency rules were implemented. Decision guidance tools were also developed to assist staff using the AV system. The mean proportion of tests and samples autoverified improved from <62% for samples and <80% for tests, to >90% for samples and >95% for tests across all three sites. The new AV system significantly decreased turn-around time and total sample review time (to about a third), however, time spent for manual review of held samples almost tripled. There was no evidence of compromise to the quality of testing process and <1% of samples held for exceeding delta check or extreme limits required corrective action. The Six Sigma (DMAIC) process improvement methodology was successfully applied to AV systems resulting in an increase in overall test and sample AV by >90%, improved turn-around time, reduced time for manual verification, and with no obvious compromise to quality or error detection. Copyright © 2018 The Canadian Society of Clinical Chemists. Published by Elsevier Inc. All rights reserved.
Testing of a Microfluidic Sampling System for High Temperature Electrochemical MC&A
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pereira, Candido; Nichols, Kevin
2013-11-27
This report describes the preliminary validation of a high-temperature microfluidic chip system for sampling of electrochemical process salt. Electroanalytical and spectroscopic techniques are attractive candidates for improvement through high-throughput sample analysis via miniaturization. Further, microfluidic chip systems are amenable to micro-scale chemical processing such as rapid, automated sample purification to improve sensor performance. The microfluidic chip was tested to determine the feasibility of the system for high temperature applications and conditions under which microfluidic systems can be used to generate salt droplets at process temperature to support development of material balance and control systems in a used fuel treatment facility.more » In FY13, the project focused on testing a quartz microchip device with molten salts at near process temperatures. The equipment was installed in glove box and tested up to 400°C using commercial thermal transfer fluids as the carrier phase. Preliminary tests were carried out with a low-melting halide salt to initially characterize the properties of this novel liquid-liquid system and to investigate the operating regimes for inducing droplet flow within candidate carrier fluids. Initial results show that the concept is viable for high temperature sampling but further development is required to optimize the system to operate with process relevant molten salts.« less
Plasma heating for containerless and microgravity materials processing
NASA Technical Reports Server (NTRS)
Leung, Emily W. (Inventor); Man, Kin F. (Inventor)
1994-01-01
A method for plasma heating of levitated samples to be used in containerless microgravity processing is disclosed. A sample is levitated by electrostatic, electromagnetic, aerodynamic, or acoustic systems, as is appropriate for the physical properties of the particular sample. The sample is heated by a plasma torch at atmospheric pressure. A ground plate is provided to help direct the plasma towards the sample. In addition, Helmholtz coils are provided to produce a magnetic field that can be used to spiral the plasma around the sample. The plasma heating system is oriented such that it does not interfere with the levitation system.
An Automated Sample Processing System for Planetary Exploration
NASA Technical Reports Server (NTRS)
Soto, Juancarlos; Lasnik, James; Roark, Shane; Beegle, Luther
2012-01-01
An Automated Sample Processing System (ASPS) for wet chemistry processing of organic materials on the surface of Mars has been jointly developed by Ball Aerospace and the Jet Propulsion Laboratory. The mechanism has been built and tested to demonstrate TRL level 4. This paper describes the function of the system, mechanism design, lessons learned, and several challenges that were overcome.
Fong, Erika J.; Huang, Chao; Hamilton, Julie; ...
2015-11-23
Here, a major advantage of microfluidic devices is the ability to manipulate small sample volumes, thus reducing reagent waste and preserving precious sample. However, to achieve robust sample manipulation it is necessary to address device integration with the macroscale environment. To realize repeatable, sensitive particle separation with microfluidic devices, this protocol presents a complete automated and integrated microfluidic platform that enables precise processing of 0.15–1.5 ml samples using microfluidic devices. Important aspects of this system include modular device layout and robust fixtures resulting in reliable and flexible world to chip connections, and fully-automated fluid handling which accomplishes closed-loop sample collection,more » system cleaning and priming steps to ensure repeatable operation. Different microfluidic devices can be used interchangeably with this architecture. Here we incorporate an acoustofluidic device, detail its characterization, performance optimization, and demonstrate its use for size-separation of biological samples. By using real-time feedback during separation experiments, sample collection is optimized to conserve and concentrate sample. Although requiring the integration of multiple pieces of equipment, advantages of this architecture include the ability to process unknown samples with no additional system optimization, ease of device replacement, and precise, robust sample processing.« less
21 CFR 111.80 - What representative samples must you collect?
Code of Federal Regulations, 2010 CFR
2010-04-01
... Process Control System § 111.80 What representative samples must you collect? The representative samples... unique lot within each unique shipment); (b) Representative samples of in-process materials for each manufactured batch at points, steps, or stages, in the manufacturing process as specified in the master...
Aircraft adaptive learning control
NASA Technical Reports Server (NTRS)
Lee, P. S. T.; Vanlandingham, H. F.
1979-01-01
The optimal control theory of stochastic linear systems is discussed in terms of the advantages of distributed-control systems, and the control of randomly-sampled systems. An optimal solution to longitudinal control is derived and applied to the F-8 DFBW aircraft. A randomly-sampled linear process model with additive process and noise is developed.
ERIC Educational Resources Information Center
Weinberger, Elizabeth
The document contains optical scannable forms for some of the instruments in the Input and Process Batteries, and guidelines for administration of the instruments in the Input Batteries of the Management Information System for Occupational Education (MISOE) Sample Data Systems. Input information describes the characteristics of the students at…
40 CFR 265.1055 - Standards: Sampling connection systems.
Code of Federal Regulations, 2013 CFR
2013-07-01
... with the requirements of § 265.1060 of this subpart. (c) In-situ sampling systems and sampling systems... required in paragraph (a) of this section shall: (1) Return the purged process fluid directly to the...
40 CFR 265.1055 - Standards: Sampling connection systems.
Code of Federal Regulations, 2014 CFR
2014-07-01
... with the requirements of § 265.1060 of this subpart. (c) In-situ sampling systems and sampling systems... required in paragraph (a) of this section shall: (1) Return the purged process fluid directly to the...
40 CFR 265.1055 - Standards: Sampling connection systems.
Code of Federal Regulations, 2012 CFR
2012-07-01
... with the requirements of § 265.1060 of this subpart. (c) In-situ sampling systems and sampling systems... required in paragraph (a) of this section shall: (1) Return the purged process fluid directly to the...
USDA-ARS?s Scientific Manuscript database
Small, coded, pill-sized tracers embedded in grain are proposed as a method for grain traceability. A sampling process for a grain traceability system was designed and investigated by applying probability statistics using a science-based sampling approach to collect an adequate number of tracers fo...
Thermal imaging measurement of lateral diffusivity and non-invasive material defect detection
Sun, Jiangang; Deemer, Chris
2003-01-01
A system and method for determining lateral thermal diffusivity of a material sample using a heat pulse; a sample oriented within an orthogonal coordinate system; an infrared camera; and a computer that has a digital frame grabber, and data acquisition and processing software. The mathematical model used within the data processing software is capable of determining the lateral thermal diffusivity of a sample of finite boundaries. The system and method may also be used as a nondestructive method for detecting and locating cracks within the material sample.
Environmental sampling can be difficult and expensive to carry out. Those taking the samples would like to integrate their knowledge of the system of study or their judgment about the system into the sample selection process to decrease the number of necessary samples. However,...
Methods and systems for detection of radionuclides
Coates, Jr., John T.; DeVol, Timothy A.
2010-05-25
Disclosed are materials and systems useful in determining the existence of radionuclides in an aqueous sample. The materials provide the dual function of both extraction and scintillation to the systems. The systems can be both portable and simple to use, and as such can beneficially be utilized to determine presence and optionally concentration of radionuclide contamination in an aqueous sample at any desired location and according to a relatively simple process without the necessity of complicated sample handling techniques. The disclosed systems include a one-step process, providing simultaneous extraction and detection capability, and a two-step process, providing a first extraction step that can be carried out in a remote field location, followed by a second detection step that can be carried out in a different location.
Understanding scaling through history-dependent processes with collapsing sample space.
Corominas-Murtra, Bernat; Hanel, Rudolf; Thurner, Stefan
2015-04-28
History-dependent processes are ubiquitous in natural and social systems. Many such stochastic processes, especially those that are associated with complex systems, become more constrained as they unfold, meaning that their sample space, or their set of possible outcomes, reduces as they age. We demonstrate that these sample-space-reducing (SSR) processes necessarily lead to Zipf's law in the rank distributions of their outcomes. We show that by adding noise to SSR processes the corresponding rank distributions remain exact power laws, p(x) ~ x(-λ), where the exponent directly corresponds to the mixing ratio of the SSR process and noise. This allows us to give a precise meaning to the scaling exponent in terms of the degree to which a given process reduces its sample space as it unfolds. Noisy SSR processes further allow us to explain a wide range of scaling exponents in frequency distributions ranging from α = 2 to ∞. We discuss several applications showing how SSR processes can be used to understand Zipf's law in word frequencies, and how they are related to diffusion processes in directed networks, or aging processes such as in fragmentation processes. SSR processes provide a new alternative to understand the origin of scaling in complex systems without the recourse to multiplicative, preferential, or self-organized critical processes.
Process observation in fiber laser-based selective laser melting
NASA Astrophysics Data System (ADS)
Thombansen, Ulrich; Gatej, Alexander; Pereira, Milton
2015-01-01
The process observation in selective laser melting (SLM) focuses on observing the interaction point where the powder is processed. To provide process relevant information, signals have to be acquired that are resolved in both time and space. Especially in high-power SLM, where more than 1 kW of laser power is used, processing speeds of several meters per second are required for a high-quality processing results. Therefore, an implementation of a suitable process observation system has to acquire a large amount of spatially resolved data at low sampling speeds or it has to restrict the acquisition to a predefined area at a high sampling speed. In any case, it is vitally important to synchronously record the laser beam position and the acquired signal. This is a prerequisite that allows the recorded data become information. Today, most SLM systems employ f-theta lenses to focus the processing laser beam onto the powder bed. This report describes the drawbacks that result for process observation and suggests a variable retro-focus system which solves these issues. The beam quality of fiber lasers delivers the processing laser beam to the powder bed at relevant focus diameters, which is a key prerequisite for this solution to be viable. The optical train we present here couples the processing laser beam and the process observation coaxially, ensuring consistent alignment of interaction zone and observed area. With respect to signal processing, we have developed a solution that synchronously acquires signals from a pyrometer and the position of the laser beam by sampling the data with a field programmable gate array. The relevance of the acquired signals has been validated by the scanning of a sample filament. Experiments with grooved samples show a correlation between different powder thicknesses and the acquired signals at relevant processing parameters. This basic work takes a first step toward self-optimization of the manufacturing process in SLM. It enables the addition of cognitive functions to the manufacturing system to the extent that the system could track its own process. The results are based on analyzing and redesigning the optical train, in combination with a real-time signal acquisition system which provides a solution to certain technological barriers.
Gumus, Abdurrahman; Ahsan, Syed; Dogan, Belgin; Jiang, Li; Snodgrass, Ryan; Gardner, Andrea; Lu, Zhengda; Simpson, Kenneth; Erickson, David
2016-01-01
The use of point-of-care (POC) devices in limited resource settings where access to commonly used infrastructure, such as water and electricity, can be restricted represents simultaneously one of the best application fits for POC systems as well as one of the most challenging places to deploy them. Of the many challenges involved in these systems, the preparation and processing of complex samples like stool, vomit, and biopsies are particularly difficult due to the high number and varied nature of mechanical and chemical interferents present in the sample. Previously we have demonstrated the ability to use solar-thermal energy to perform PCR based nucleic acid amplifications. In this work demonstrate how the technique, using similar infrastructure, can also be used to perform solar-thermal based sample processing system for extracting and isolating Vibrio Cholerae nucleic acids from fecal samples. The use of opto-thermal energy enables the use of sunlight to drive thermal lysing reactions in large volumes without the need for external electrical power. Using the system demonstrate the ability to reach a 95°C threshold in less than 5 minutes and maintain a stable sample temperature of +/− 2°C following the ramp up. The system is demonstrated to provide linear results between 104 and 108 CFU/mL when the released nucleic acids were quantified via traditional means. Additionally, we couple the sample processing unit with our previously demonstrated solar-thermal PCR and tablet based detection system to demonstrate very low power sample-in-answer-out detection. PMID:27231636
NASA Astrophysics Data System (ADS)
Alanis, Elvio; Romero, Graciela; Alvarez, Liliana; Martinez, Carlos C.; Hoyos, Daniel; Basombrio, Miguel A.
2001-08-01
A fully automated image processing system for detection of motile microorganism is biological samples is presented. The system is specifically calibrated for determining the concentration of Trypanosoma Cruzi parasites in blood samples of mice infected with Chagas disease. The method can be adapted for use in other biological samples. A thin layer of blood infected by T. cruzi parasites is examined in a common microscope in which the images of the vision field are taken by a CCD camera and temporarily stored in the computer memory. In a typical field, a few motile parasites are observable surrounded by blood red cells. The parasites have low contrast. Thus, they are difficult to detect visually but their great motility betrays their presence by the movement of the nearest neighbor red cells. Several consecutive images of the same field are taken, decorrelated with each other where parasites are present, and digitally processed in order to measure the number of parasites present in the field. Several fields are sequentially processed in the same fashion, displacing the sample by means of step motors driven by the computer. A direct advantage of this system is that its results are more reliable and the process is less time consuming than the current subjective evaluations made visually by technicians.
Process test plan, phase II: waste retrieval sluicing system emissions collection
DOE Office of Scientific and Technical Information (OSTI.GOV)
POWERS, R.L.
1999-06-01
This Process Test Plan is prepared to continue from HNF-3733 which was Phase I of the test. Supplemental operational controls and sampling requirements are defined to safely obtain gas samples from the 296-C-006 ventilation system stack during active operation of the sluicing equipment.
Under-sampling in a Multiple-Channel Laser Vibrometry System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Corey, Jordan
2007-03-01
Laser vibrometry is a technique used to detect vibrations on objects using the interference of coherent light with itself. Most vibrometry systems process only one target location at a time, but processing multiple locations simultaneously provides improved detection capabilities. Traditional laser vibrometry systems employ oversampling to sample the incoming modulated-light signal, however as the number of channels increases in these systems, certain issues arise such a higher computational cost, excessive heat, increased power requirements, and increased component cost. This thesis describes a novel approach to laser vibrometry that utilizes undersampling to control the undesirable issues associated with over-sampled systems. Undersamplingmore » allows for significantly less samples to represent the modulated-light signals, which offers several advantages in the overall system design. These advantages include an improvement in thermal efficiency, lower processing requirements, and a higher immunity to the relative intensity noise inherent in laser vibrometry applications. A unique feature of this implementation is the use of a parallel architecture to increase the overall system throughput. This parallelism is realized using a hierarchical multi-channel architecture based on off-the-shelf programmable logic devices (PLDs).« less
Dahling, Daniel R
2002-01-01
Large-scale virus studies of groundwater systems require practical and sensitive procedures for both sample processing and viral assay. Filter adsorption-elution procedures have traditionally been used to process large-volume water samples for viruses. In this study, five filter elution procedures using cartridge filters were evaluated for their effectiveness in processing samples. Of the five procedures tested, the third method, which incorporated two separate beef extract elutions (one being an overnight filter immersion in beef extract), recovered 95% of seeded poliovirus compared with recoveries of 36 to 70% for the other methods. For viral enumeration, an expanded roller bottle quantal assay was evaluated using seeded poliovirus. This cytopathic-based method was considerably more sensitive than the standard plaque assay method. The roller bottle system was more economical than the plaque assay for the evaluation of comparable samples. Using roller bottles required less time and manipulation than the plaque procedure and greatly facilitated the examination of large numbers of samples. The combination of the improved filter elution procedure and the roller bottle assay for viral analysis makes large-scale virus studies of groundwater systems practical. This procedure was subsequently field tested during a groundwater study in which large-volume samples (exceeding 800 L) were processed through the filters.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fong, Erika J.; Huang, Chao; Hamilton, Julie
Here, a major advantage of microfluidic devices is the ability to manipulate small sample volumes, thus reducing reagent waste and preserving precious sample. However, to achieve robust sample manipulation it is necessary to address device integration with the macroscale environment. To realize repeatable, sensitive particle separation with microfluidic devices, this protocol presents a complete automated and integrated microfluidic platform that enables precise processing of 0.15–1.5 ml samples using microfluidic devices. Important aspects of this system include modular device layout and robust fixtures resulting in reliable and flexible world to chip connections, and fully-automated fluid handling which accomplishes closed-loop sample collection,more » system cleaning and priming steps to ensure repeatable operation. Different microfluidic devices can be used interchangeably with this architecture. Here we incorporate an acoustofluidic device, detail its characterization, performance optimization, and demonstrate its use for size-separation of biological samples. By using real-time feedback during separation experiments, sample collection is optimized to conserve and concentrate sample. Although requiring the integration of multiple pieces of equipment, advantages of this architecture include the ability to process unknown samples with no additional system optimization, ease of device replacement, and precise, robust sample processing.« less
Patel, Darshan C; Lyu, Yaqi Fara; Gandarilla, Jorge; Doherty, Steve
2018-04-03
In-process sampling and analysis is an important aspect of monitoring kinetic profiles and impurity formation or rejection, both in development and during commercial manufacturing. In pharmaceutical process development, the technology of choice for a substantial portion of this analysis is high-performance liquid chromatography (HPLC). Traditionally, the sample extraction and preparation for reaction characterization have been performed manually. This can be time consuming, laborious, and impractical for long processes. Depending on the complexity of the sample preparation, there can be variability introduced by different analysts, and in some cases, the integrity of the sample can be compromised during handling. While there are commercial instruments available for on-line monitoring with HPLC, they lack capabilities in many key areas. Some do not provide integration of the sampling and analysis, while others afford limited flexibility in sample preparation. The current offerings provide a limited number of unit operations available for sample processing and no option for workflow customizability. This work describes development of a microfluidic automated program (MAP) which fully automates the sample extraction, manipulation, and on-line LC analysis. The flexible system is controlled using an intuitive Microsoft Excel based user interface. The autonomous system is capable of unattended reaction monitoring that allows flexible unit operations and workflow customization to enable complex operations and on-line sample preparation. The automated system is shown to offer advantages over manual approaches in key areas while providing consistent and reproducible in-process data. Copyright © 2017 Elsevier B.V. All rights reserved.
Use the Bar Code System to Improve Accuracy of the Patient and Sample Identification.
Chuang, Shu-Hsia; Yeh, Huy-Pzu; Chi, Kun-Hung; Ku, Hsueh-Chen
2018-01-01
In time and correct sample collection were highly related to patient's safety. The sample error rate was 11.1%, because misbranded patient information and wrong sample containers during January to April, 2016. We developed a barcode system of "Specimens Identify System" through process of reengineering of TRM, used bar code scanners, add sample container instructions, and mobile APP. Conclusion, the bar code systems improved the patient safety and created green environment.
NASA Technical Reports Server (NTRS)
Jandura, Louise
2010-01-01
The Sample Acquisition/Sample Processing and Handling subsystem for the Mars Science Laboratory is a highly-mechanized, Rover-based sampling system that acquires powdered rock and regolith samples from the Martian surface, sorts the samples into fine particles through sieving, and delivers small portions of the powder into two science instruments inside the Rover. SA/SPaH utilizes 17 actuated degrees-of-freedom to perform the functions needed to produce 5 sample pathways in support of the scientific investigation on Mars. Both hardware redundancy and functional redundancy are employed in configuring this sampling system so some functionality is retained even with the loss of a degree-of-freedom. Intentional dynamic environments are created to move sample while vibration isolators attenuate this environment at the sensitive instruments located near the dynamic sources. In addition to the typical flight hardware qualification test program, two additional types of testing are essential for this kind of sampling system: characterization of the intentionally-created dynamic environment and testing of the sample acquisition and processing hardware functions using Mars analog materials in a low pressure environment. The overall subsystem design and configuration are discussed along with some of the challenges, tradeoffs, and lessons learned in the areas of fault tolerance, intentional dynamic environments, and special testing
Robotic sampling system for an unmanned Mars mission
NASA Technical Reports Server (NTRS)
Chun, Wendell
1989-01-01
A major robotics opportunity for NASA will be the Mars Rover/Sample Return Mission which could be launched as early as the 1990s. The exploratory portion of this mission will include two autonomous subsystems: the rover vehicle and a sample handling system. The sample handling system is the key to the process of collecting Martian soils. This system could include a core drill, a general-purpose manipulator, tools, containers, a return canister, certification hardware and a labeling system. Integrated into a functional package, the sample handling system is analogous to a complex robotic workcell. Discussed here are the different components of the system, their interfaces, forseeable problem areas and many options based on the scientific goals of the mission. The various interfaces in the sample handling process (component to component and handling system to rover) will be a major engineering effort. Two critical evaluation criteria that will be imposed on the system are flexibility and reliability. It needs to be flexible enough to adapt to different scenarios and environments and acquire the most desirable specimens for return to Earth. Scientists may decide to change the distribution and ratio of core samples to rock samples in the canister. The long distance and duration of this planetary mission places a reliability burden on the hardware. The communication time delay between Earth and Mars minimizes operator interaction (teleoperation, supervisory modes) with the sample handler. An intelligent system will be required to plan the actions, make sample choices, interpret sensor inputs, and query unknown surroundings. A combination of autonomous functions and supervised movements will be integrated into the sample handling system.
These Society of Toxicologic Pathology “best” practice recommendations should ensure consistent sampling, processing, and evaluation of the peripheral nervous system (PNS). For toxicity studies where neurotoxicity is not anticipated (Situation 1), PNS evaluation may be limited...
Note: A simple sample transfer alignment for ultra-high vacuum systems.
Tamtögl, A; Carter, E A; Ward, D J; Avidor, N; Kole, P R; Jardine, A P; Allison, W
2016-06-01
The alignment of ultra-high-vacuum sample transfer systems can be problematic when there is no direct line of sight to assist the user. We present the design of a simple and cheap system which greatly simplifies the alignment of sample transfer devices. Our method is based on the adaptation of a commercial digital camera which provides live views from within the vacuum chamber. The images of the camera are further processed using an image recognition and processing code which determines any misalignments and reports them to the user. Installation has proven to be extremely useful in order to align the sample with respect to the transfer mechanism. Furthermore, the alignment software can be easily adapted for other systems.
Time lens assisted photonic sampling extraction
NASA Astrophysics Data System (ADS)
Petrillo, Keith Gordon
Telecommunication bandwidth demands have dramatically increased in recent years due to Internet based services like cloud computing and storage, large file sharing, and video streaming. Additionally, sensing systems such as wideband radar, magnetic imaging resonance systems, and complex modulation formats to handle large data transfer in telecommunications require high speed, high resolution analog-to-digital converters (ADCs) to interpret the data. Accurately processing and acquiring the information at next generation data rates from these systems has become challenging for electronic systems. The largest contributors to the electronic bottleneck are bandwidth and timing jitter which limit speed and reduce accuracy. Optical systems have shown to have at least three orders of magnitude increase in bandwidth capabilities and state of the art mode locked lasers have reduced timing jitters into thousands of attoseconds. Such features have encouraged processing signals without the use of electronics or using photonics to assist electronics. All optical signal processing has allowed the processing of telecommunication line rates up to 1.28 Tb/s and high resolution analog-to-digital converters in the 10s of gigahertz. The major drawback to these optical systems is the high cost of the components. The application of all optical processing techniques such as a time lens and chirped processing can greatly reduce bandwidth and cost requirements of optical serial to parallel converters and push photonically assisted ADCs into the 100s of gigahertz. In this dissertation, the building blocks to a high speed photonically assisted ADC are demonstrated, each providing benefits to its own respective application. A serial to parallel converter using a continuously operating time lens as an optical Fourier processor is demonstrated to fully convert a 160-Gb/s optical time division multiplexed signal to 16 10-Gb/s channels with error free operation. Using chirped processing, an optical sample and hold concept is demonstrated and analyzed as a resolution improvement to existing photonically assisted ADCs. Simulations indicate that the application of a continuously operating time lens to a photonically assisted sampling system can increase photonically sampled systems by an order of magnitude while acquiring properties similar to an optical sample and hold system.
Networks for image acquisition, processing and display
NASA Technical Reports Server (NTRS)
Ahumada, Albert J., Jr.
1990-01-01
The human visual system comprises layers of networks which sample, process, and code images. Understanding these networks is a valuable means of understanding human vision and of designing autonomous vision systems based on network processing. Ames Research Center has an ongoing program to develop computational models of such networks. The models predict human performance in detection of targets and in discrimination of displayed information. In addition, the models are artificial vision systems sharing properties with biological vision that has been tuned by evolution for high performance. Properties include variable density sampling, noise immunity, multi-resolution coding, and fault-tolerance. The research stresses analysis of noise in visual networks, including sampling, photon, and processing unit noises. Specific accomplishments include: models of sampling array growth with variable density and irregularity comparable to that of the retinal cone mosaic; noise models of networks with signal-dependent and independent noise; models of network connection development for preserving spatial registration and interpolation; multi-resolution encoding models based on hexagonal arrays (HOP transform); and mathematical procedures for simplifying analysis of large networks.
Integrated microfluidic systems for cell lysis, mixing/pumping and DNA amplification
NASA Astrophysics Data System (ADS)
Lee, Chia-Yen; Lee, Gwo-Bin; Lin, Jr-Lung; Huang, Fu-Chun; Liao, Chia-Sheng
2005-06-01
The present paper reports a fully automated microfluidic system for the DNA amplification process by integrating an electroosmotic pump, an active micromixer and an on-chip temperature control system. In this DNA amplification process, the cell lysis is initially performed in a micro cell lysis reactor. Extracted DNA samples, primers and reagents are then driven electroosmotically into a mixing region where they are mixed by the active micromixer. The homogeneous mixture is then thermally cycled in a micro-PCR (polymerase chain reaction) chamber to perform DNA amplification. Experimental results show that the proposed device can successfully automate the sample pretreatment operation for DNA amplification, thereby delivering significant time and effort savings. The new microfluidic system, which facilitates cell lysis, sample driving/mixing and DNA amplification, could provide a significant contribution to ongoing efforts to miniaturize bio-analysis systems by utilizing a simple fabrication process and cheap materials.
JPRS Report, Science & Technology, Japan, High Temperature Materials
1990-11-09
3 This restriction is heavy. The inconvenience that the material powder of the imido thermal composition method, for example, which shows the best...procedures, system composition , features of formability, and forming characteristic of forming samples using alumina material will be made. 2...Osmotic V Process Forming System 2.1 System Composition of Process A system block diagram of the process is shown in Figure 1. The V process forming system
Rapid DNA analysis for automated processing and interpretation of low DNA content samples.
Turingan, Rosemary S; Vasantgadkar, Sameer; Palombo, Luke; Hogan, Catherine; Jiang, Hua; Tan, Eugene; Selden, Richard F
2016-01-01
Short tandem repeat (STR) analysis of casework samples with low DNA content include those resulting from the transfer of epithelial cells from the skin to an object (e.g., cells on a water bottle, or brim of a cap), blood spatter stains, and small bone and tissue fragments. Low DNA content (LDC) samples are important in a wide range of settings, including disaster response teams to assist in victim identification and family reunification, military operations to identify friend or foe, criminal forensics to identify suspects and exonerate the innocent, and medical examiner and coroner offices to identify missing persons. Processing LDC samples requires experienced laboratory personnel, isolated workstations, and sophisticated equipment, requires transport time, and involves complex procedures. We present a rapid DNA analysis system designed specifically to generate STR profiles from LDC samples in field-forward settings by non-technical operators. By performing STR in the field, close to the site of collection, rapid DNA analysis has the potential to increase throughput and to provide actionable information in real time. A Low DNA Content BioChipSet (LDC BCS) was developed and manufactured by injection molding. It was designed to function in the fully integrated Accelerated Nuclear DNA Equipment (ANDE) instrument previously designed for analysis of buccal swab and other high DNA content samples (Investigative Genet. 4(1):1-15, 2013). The LDC BCS performs efficient DNA purification followed by microfluidic ultrafiltration of the purified DNA, maximizing the quantity of DNA available for subsequent amplification and electrophoretic separation and detection of amplified fragments. The system demonstrates accuracy, precision, resolution, signal strength, and peak height ratios appropriate for casework analysis. The LDC rapid DNA analysis system is effective for the generation of STR profiles from a wide range of sample types. The technology broadens the range of sample types that can be processed and minimizes the time between sample collection, sample processing and analysis, and generation of actionable intelligence. The fully integrated Expert System is capable of interpreting a wide range or sample types and input DNA quantities, allowing samples to be processed and interpreted without a technical operator.
Mayer, Horst; Brümmer, Jens; Brinkmann, Thomas
2011-01-01
To implement Lean Six Sigma in our central laboratory we conducted a project to measure single pre-analytical steps influencing turnaround time (TAT) of emergency department (ED) serum samples. The traditional approach of extracting data from the Laboratory Information System (LIS) for a retrospective calculation of a mean TAT is not suitable. Therefore, we used radiofrequency identification (RFID) chips for real time tracking of individual samples at any pre-analytical step. 1,200 serum tubes were labelled with RFID chips and were provided to the emergency department. 3 RFID receivers were installed in the laboratory: at the outlet of the pneumatic tube system, at the centrifuge, and in the analyser area. In addition, time stamps of sample entry at the automated sample distributor and communication of results from the analyser were collected from LIS. 1,023 labelled serum tubes arrived at our laboratory. 899 RFID tags were used for TAT calculation. The following transfer times were determined (median 95th percentile in min:sec): pneumatic tube system --> centrifuge (01:25/04:48), centrifuge --> sample distributor (14:06/5:33), sample distributor --> analysis system zone (02:39/15:07), analysis system zone --> result communication (12:42/22:21). Total TAT was calculated at 33:19/57:40 min:sec. Manual processes around centrifugation were identified as a major part of TAT with 44%/60% (median/95th percentile). RFID is a robust, easy to use, and error-free technology and not susceptible to interferences in the laboratory environment. With this study design we were able to measure significant variations in a single manual sample transfer process. We showed that TAT is mainly influenced by manual steps around the centrifugation process and we concluded that centrifugation should be integrated in solutions for total laboratory automation.
NASA Astrophysics Data System (ADS)
Jandura, L.; Burke, K.; Kennedy, B.; Melko, J.; Okon, A.; Sunshine, D.
2009-12-01
The Sample Acquisition/Sample Processing and Handling (SA/SPaH) subsystem for the Mars Science Library (MSL) is a rover-based sampling system scheduled to launch in 2011. The SA/SPaH consists of a powdering drill and a scooping, sieving, and portioning device mounted on a turret at the end of a robotic arm. Also on the turret is a dust removal tool for clearing the surface of scientific targets, and two science instruments mounted on vibration isolators. The SA/SPaH can acquire powder from rocks at depths of 20 to 50 mm and can also pick up loose regolith with its scoop. The acquired sample is sieved and portioned and delivered to one of two instruments inside the rover for analysis. The functionality of the system will be described along with the targets the system can acquire and the sample that can be delivered. Top View of the SA/SPaH on the Rover
Improving the Acquisition and Management of Sample Curation Data
NASA Technical Reports Server (NTRS)
Todd, Nancy S.; Evans, Cindy A.; Labasse, Dan
2011-01-01
This paper discusses the current sample documentation processes used during and after a mission, examines the challenges and special considerations needed for designing effective sample curation data systems, and looks at the results of a simulated sample result mission and the lessons learned from this simulation. In addition, it introduces a new data architecture for an integrated sample Curation data system being implemented at the NASA Astromaterials Acquisition and Curation department and discusses how it improves on existing data management systems.
ERIC Educational Resources Information Center
Mitchell, Eugene E., Ed.
The simulation of a sampled-data system is described that uses a full parallel hybrid computer. The sampled data system simulated illustrates the proportional-integral-derivative (PID) discrete control of a continuous second-order process representing a stirred-tank. The stirred-tank is simulated using continuous analog components, while PID…
NASA Astrophysics Data System (ADS)
Yussup, N.; Ibrahim, M. M.; Rahman, N. A. A.; Mokhtar, M.; Salim, N. A. A.; Soh@Shaari, S. C.; Azman, A.; Lombigit, L.; Azman, A.; Omar, S. A.
2018-01-01
Most of the procedures in neutron activation analysis (NAA) process that has been established in Malaysian Nuclear Agency (Nuclear Malaysia) since 1980s were performed manually. These manual procedures carried out by the NAA laboratory personnel are time consuming and inefficient especially for sample counting and measurement process. The sample needs to be changed and the measurement software needs to be setup for every one hour counting time. Both of these procedures are performed manually for every sample. Hence, an automatic sample changer system (ASC) that consists of hardware and software is developed to automate sample counting process for up to 30 samples consecutively. This paper describes the ASC control software for NAA process which is designed and developed to control the ASC hardware and call GammaVision software for sample measurement. The software is developed by using National Instrument LabVIEW development package.
Pohl, Fabian; Hartmann, Werner; Holzmann, Thomas; Gensicke, Sandra; Kölbl, Oliver; Hautmann, Matthias G
2014-01-25
Many cancer patients receive a central venous catheter or port system prior to therapy to assure correct drug administration. Even appropriate hygienic intervention maintenance carries the risk of contaminating the middle port (C-port) of a three-way cock (TWC), a risk that increases with the number of medical interventions. Because of the complexity of the cleaning procedure with disconnection and reconnection of the standard luer lock cap (referred as "intervention"), we compared luer lock caps with a "closed access system" consisting of a luer access split septum system with regard to process optimization (work simplification, process time), efficiency (costs) and hygiene (patient safety). For determination of process optimization the workflow of an intervention according to the usual practice and risks was depicted in a process diagram. For determining the actual process costs, we analyzed use of material and time parameters per intervention and used the process parameters for programming the process into a simulation run (n = 1000) to determine the process costs as well as their differences (ACTUAL vs. NOMINAL) within the framework of a discrete event simulation.Additionally cultures were carried out at the TWC C-ports to evaluate possible contamination. With the closed access system, the mean working time of 5.5 minutes could be reduced to 2.97 minutes. The results for average process costs (labour and material costs per use) were 3.92 € for luer lock caps and 2.55 € for the closed access system. The hypothesis test (2-sample t-test, CI 0.95, p-value<0.05) confirmed the significance of the result.In 50 reviewed samples (TWC's), the contamination rate for the luer lock cap was 8% (4 out of 50 samples were positive), the contamination rate of the 50 samples with the closed access system was 0%.Possible hygienic risks (related to material, surroundings, staff handling) could be reduced by 65.38%. In the present research, the closed access system with a divided split septum was superior to conventional luer lock caps. The advantage of the closed access system lies in the simplified handling for staff, which results in a reduced risk of patient infection due to improved clinical hygiene.
Design and fabrication of a glovebox for the Plasma Hearth Process radioactive bench-scale system
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wahlquist, D.R.
This paper presents some of the design considerations and fabrication techniques for building a glovebox for the Plasma Hearth Process (PHP) radioactive bench-scale system. The PHP radioactive bench-scale system uses a plasma torch to process a variety of radioactive materials into a final vitrified waste form. The processed waste will contain plutonium and trace amounts of other radioactive materials. The glovebox used in this system is located directly below the plasma chamber and is called the Hearth Handling Enclosure (HHE). The HHE is designed to maintain a confinement boundary between the processed waste and the operator. Operations that take placemore » inside the HHE include raising and lowering the hearth using a hydraulic lift table, transporting the hearth within the HHE using an overhead monorail and hoist system, sampling and disassembly of the processed waste and hearth, weighing the hearth, rebuilding a hearth, and sampling HEPA filters. The PHP radioactive bench-scale system is located at the TREAT facility at Argonne National Laboratory-West in Idaho Falls, Idaho.« less
A robust variable sampling time BLDC motor control design based upon μ-synthesis.
Hung, Chung-Wen; Yen, Jia-Yush
2013-01-01
The variable sampling rate system is encountered in many applications. When the speed information is derived from the position marks along the trajectory, one would have a speed dependent sampling rate system. The conventional fixed or multisampling rate system theory may not work in these cases because the system dynamics include the uncertainties which resulted from the variable sampling rate. This paper derived a convenient expression for the speed dependent sampling rate system. The varying sampling rate effect is then translated into multiplicative uncertainties to the system. The design then uses the popular μ-synthesis process to achieve a robust performance controller design. The implementation on a BLDC motor demonstrates the effectiveness of the design approach.
A Robust Variable Sampling Time BLDC Motor Control Design Based upon μ-Synthesis
Yen, Jia-Yush
2013-01-01
The variable sampling rate system is encountered in many applications. When the speed information is derived from the position marks along the trajectory, one would have a speed dependent sampling rate system. The conventional fixed or multisampling rate system theory may not work in these cases because the system dynamics include the uncertainties which resulted from the variable sampling rate. This paper derived a convenient expression for the speed dependent sampling rate system. The varying sampling rate effect is then translated into multiplicative uncertainties to the system. The design then uses the popular μ-synthesis process to achieve a robust performance controller design. The implementation on a BLDC motor demonstrates the effectiveness of the design approach. PMID:24327804
Ren, Luquan; Zhou, Xueli; Song, Zhengyi; Zhao, Che; Liu, Qingping; Xue, Jingze; Li, Xiujuan
2017-03-16
Recently, with a broadening range of available materials and alteration of feeding processes, several extrusion-based 3D printing processes for metal materials have been developed. An emerging process is applicable for the fabrication of metal parts into electronics and composites. In this paper, some critical parameters of extrusion-based 3D printing processes were optimized by a series of experiments with a melting extrusion printer. The raw materials were copper powder and a thermoplastic organic binder system and the system included paraffin wax, low density polyethylene, and stearic acid (PW-LDPE-SA). The homogeneity and rheological behaviour of the raw materials, the strength of the green samples, and the hardness of the sintered samples were investigated. Moreover, the printing and sintering parameters were optimized with an orthogonal design method. The influence factors in regard to the ultimate tensile strength of the green samples can be described as follows: infill degree > raster angle > layer thickness. As for the sintering process, the major factor on hardness is sintering temperature, followed by holding time and heating rate. The highest hardness of the sintered samples was very close to the average hardness of commercially pure copper material. Generally, the extrusion-based printing process for producing metal materials is a promising strategy because it has some advantages over traditional approaches for cost, efficiency, and simplicity.
Ren, Luquan; Zhou, Xueli; Song, Zhengyi; Zhao, Che; Liu, Qingping; Xue, Jingze; Li, Xiujuan
2017-01-01
Recently, with a broadening range of available materials and alteration of feeding processes, several extrusion-based 3D printing processes for metal materials have been developed. An emerging process is applicable for the fabrication of metal parts into electronics and composites. In this paper, some critical parameters of extrusion-based 3D printing processes were optimized by a series of experiments with a melting extrusion printer. The raw materials were copper powder and a thermoplastic organic binder system and the system included paraffin wax, low density polyethylene, and stearic acid (PW–LDPE–SA). The homogeneity and rheological behaviour of the raw materials, the strength of the green samples, and the hardness of the sintered samples were investigated. Moreover, the printing and sintering parameters were optimized with an orthogonal design method. The influence factors in regard to the ultimate tensile strength of the green samples can be described as follows: infill degree > raster angle > layer thickness. As for the sintering process, the major factor on hardness is sintering temperature, followed by holding time and heating rate. The highest hardness of the sintered samples was very close to the average hardness of commercially pure copper material. Generally, the extrusion-based printing process for producing metal materials is a promising strategy because it has some advantages over traditional approaches for cost, efficiency, and simplicity. PMID:28772665
Vorberg, Ellen; Fleischer, Heidi; Junginger, Steffen; Liu, Hui; Stoll, Norbert; Thurow, Kerstin
2016-10-01
Life science areas require specific sample pretreatment to increase the concentration of the analytes and/or to convert the analytes into an appropriate form for the detection and separation systems. Various workstations are commercially available, allowing for automated biological sample pretreatment. Nevertheless, due to the required temperature, pressure, and volume conditions in typical element and structure-specific measurements, automated platforms are not suitable for analytical processes. Thus, the purpose of the presented investigation was the design, realization, and evaluation of an automated system ensuring high-precision sample preparation for a variety of analytical measurements. The developed system has to enable system adaption and high performance flexibility. Furthermore, the system has to be capable of dealing with the wide range of required vessels simultaneously, allowing for less cost and time-consuming process steps. However, the system's functionality has been confirmed in various validation sequences. Using element-specific measurements, the automated system was up to 25% more precise compared to the manual procedure and as precise as the manual procedure using structure-specific measurements. © 2015 Society for Laboratory Automation and Screening.
NASA Astrophysics Data System (ADS)
Knight, Travis Warren
Nuclear thermal propulsion (NTP) and space nuclear power are two enabling technologies for the manned exploration of space and the development of research outposts in space and on other planets such as Mars. Advanced carbide nuclear fuels have been proposed for application in space nuclear power and propulsion systems. This study examined the processing technologies and optimal parameters necessary to fabricate samples of single phase, solid solution, mixed uranium/refractory metal carbides. In particular, the pseudo-ternary carbide, UC-ZrC-NbC, system was examined with uranium metal mole fractions of 5% and 10% and corresponding uranium densities of 0.8 to 1.8 gU/cc. Efforts were directed to those methods that could produce simple geometry fuel elements or wafers such as those used to fabricate a Square Lattice Honeycomb (SLHC) fuel element and reactor core. Methods of cold uniaxial pressing, sintering by induction heating, and hot pressing by self-resistance heating were investigated. Solid solution, high density (low porosity) samples greater than 95% TD were processed by cold pressing at 150 MPa and sintering above 2600 K for times longer than 90 min. Some impurity oxide phases were noted in some samples attributed to residual gases in the furnace during processing. Also, some samples noted secondary phases of carbon and UC2 due to some hyperstoichiometric powder mixtures having carbon-to-metal ratios greater than one. In all, 33 mixed carbide samples were processed and analyzed with half bearing uranium as ternary carbides of UC-ZrC-NbC. Scanning electron microscopy, x-ray diffraction, and density measurements were used to characterize samples. Samples were processed from powders of the refractory mono-carbides and UC/UC 2 or from powders of uranium hydride (UH3), graphite, and refractory metal carbides to produce hypostoichiometric mixed carbides. Samples processed from the constituent carbide powders and sintered at temperatures above the melting point of UC showed signs of liquid phase sintering and were shown to be largely solid solutions. Pre-compaction of mixed carbide powders prior to sintering was shown to be necessary to achieve high densities. Hypostoichiometric, samples processed at 2500 K exhibited only the initial stage of sintering and solid solution formation. Based on these findings, a suggested processing methodology is proposed for producing high density, solid solution, mixed carbide fuels. Pseudo-binary, refractory carbide samples hot pressed at 3100 K and 6 MPa showed comparable densities (approximately 85% of the theoretical value) to samples processed by cold pressing and sintering at temperatures of 2800 K.
Remote Sensing Image Quality Assessment Experiment with Post-Processing
NASA Astrophysics Data System (ADS)
Jiang, W.; Chen, S.; Wang, X.; Huang, Q.; Shi, H.; Man, Y.
2018-04-01
This paper briefly describes the post-processing influence assessment experiment, the experiment includes three steps: the physical simulation, image processing, and image quality assessment. The physical simulation models sampled imaging system in laboratory, the imaging system parameters are tested, the digital image serving as image processing input are produced by this imaging system with the same imaging system parameters. The gathered optical sampled images with the tested imaging parameters are processed by 3 digital image processes, including calibration pre-processing, lossy compression with different compression ratio and image post-processing with different core. Image quality assessment method used is just noticeable difference (JND) subject assessment based on ISO20462, through subject assessment of the gathered and processing images, the influence of different imaging parameters and post-processing to image quality can be found. The six JND subject assessment experimental data can be validated each other. Main conclusions include: image post-processing can improve image quality; image post-processing can improve image quality even with lossy compression, image quality with higher compression ratio improves less than lower ratio; with our image post-processing method, image quality is better, when camera MTF being within a small range.
Study and Analysis of The Robot-Operated Material Processing Systems (ROMPS)
NASA Technical Reports Server (NTRS)
Nguyen, Charles C.
1996-01-01
This is a report presenting the progress of a research grant funded by NASA for work performed during 1 Oct. 1994 - 31 Sep. 1995. The report deals with the development and investigation of potential use of software for data processing for the Robot Operated Material Processing System (ROMPS). It reports on the progress of data processing of calibration samples processed by ROMPS in space and on earth. First data were retrieved using the I/O software and manually processed using MicroSoft Excel. Then the data retrieval and processing process was automated using a program written in C which is able to read the telemetry data and produce plots of time responses of sample temperatures and other desired variables. LabView was also employed to automatically retrieve and process the telemetry data.
General purpose rocket furnace
NASA Technical Reports Server (NTRS)
Aldrich, B. R.; Whitt, W. D. (Inventor)
1979-01-01
A multipurpose furnace for space vehicles used for material processing experiments in an outer space environment is described. The furnace contains three separate cavities designed to process samples of the widest possible range of materials and thermal requirements. Each cavity contains three heating elements capable of independent function under the direction of an automatic and programmable control system. A heat removable mechanism is also provided for each cavity which operates in conjunction with the control system for establishing an isothermally heated cavity or a wide range of thermal gradients and cool down rates. A monitoring system compatible with the rocket telemetry provides furnace performance and sample growth rate data throughout the processing cycle.
Topin, Sylvain; Greau, Claire; Deliere, Ludovic; Hovesepian, Alexandre; Taffary, Thomas; Le Petit, Gilbert; Douysset, Guilhem; Moulin, Christophe
2015-11-01
The SPALAX (Système de Prélèvement Automatique en Ligne avec l'Analyse du Xénon) is one of the systems used in the International Monitoring System of the Comprehensive Nuclear Test Ban Treaty (CTBT) to detect radioactive xenon releases following a nuclear explosion. Approximately 10 years after the industrialization of the first system, the CEA has developed the SPALAX New Generation, SPALAX-NG, with the aim of increasing the global sensitivity and reducing the overall size of the system. A major breakthrough has been obtained by improving the sampling stage and the purification/concentration stage. The sampling stage evolution consists of increasing the sampling capacity and improving the gas treatment efficiency across new permeation membranes, leading to an increase in the xenon production capacity by a factor of 2-3. The purification/concentration stage evolution consists of using a new adsorbent Ag@ZSM-5 (or Ag-PZ2-25) with a much larger xenon retention capacity than activated charcoal, enabling a significant reduction in the overall size of this stage. The energy consumption of the system is similar to that of the current SPALAX system. The SPALAX-NG process is able to produce samples of almost 7 cm(3) of xenon every 12 h, making it the most productive xenon process among the IMS systems. Copyright © 2015 Elsevier Ltd. All rights reserved.
2007-03-01
Characterisation. In Nanotechnology Aerospace Applications – 2006 (pp. 4-1 – 4-8). Educational Notes RTO-EN-AVT-129bis, Paper 4. Neuilly-sur-Seine, France: RTO...the Commercialisation Processes Concept IDEA Proof-of- Principle Trial Samples Engineering Verification Samples Design Verification Samples...SEIC Systems Engineering for commercialisation Design Houses, Engineering & R&D USERS & Integrators SE S U R Integrators Fabs & Wafer Processing Die
Development of techniques for processing metal-metal oxide systems
NASA Technical Reports Server (NTRS)
Johnson, P. C.
1976-01-01
Techniques for producing model metal-metal oxide systems for the purpose of evaluating the results of processing such systems in the low-gravity environment afforded by a drop tower facility are described. Because of the lack of success in producing suitable materials samples and techniques for processing in the 3.5 seconds available, the program was discontinued.
Two-Stage Variable Sample-Rate Conversion System
NASA Technical Reports Server (NTRS)
Tkacenko, Andre
2009-01-01
A two-stage variable sample-rate conversion (SRC) system has been pro posed as part of a digital signal-processing system in a digital com munication radio receiver that utilizes a variety of data rates. The proposed system would be used as an interface between (1) an analog- todigital converter used in the front end of the receiver to sample an intermediatefrequency signal at a fixed input rate and (2) digita lly implemented tracking loops in subsequent stages that operate at v arious sample rates that are generally lower than the input sample r ate. This Two-Stage System would be capable of converting from an input sample rate to a desired lower output sample rate that could be var iable and not necessarily a rational fraction of the input rate.
An overview of the thematic mapper geometric correction system
NASA Technical Reports Server (NTRS)
Beyer, E. P.
1983-01-01
Geometric accuracy specifications for LANDSAT 4 are reviewed and the processing concepts which form the basis of NASA's thematic mapper geometric correction system are summarized for both the flight and ground segments. The flight segment includes the thematic mapper instrument, attitude measurement devices, attitude control, and ephemeris processing. For geometric correction the ground segment uses mirror scan correction data, payload correction data, and control point information to determine where TM detector samples fall on output map projection systems. Then the raw imagery is reformatted and resampled to produce image samples on a selected output projection grid system.
Particle-sampling statistics in laser anemometers Sample-and-hold systems and saturable systems
NASA Technical Reports Server (NTRS)
Edwards, R. V.; Jensen, A. S.
1983-01-01
The effect of the data-processing system on the particle statistics obtained with laser anemometry of flows containing suspended particles is examined. Attention is given to the sample and hold processor, a pseudo-analog device which retains the last measurement until a new measurement is made, followed by time-averaging of the data. The second system considered features a dead time, i.e., a saturable system with a significant reset time with storage in a data buffer. It is noted that the saturable system operates independent of the particle arrival rate. The probabilities of a particle arrival in a given time period are calculated for both processing systems. It is shown that the system outputs are dependent on the mean particle flow rate, the flow correlation time, and the flow statistics, indicating that the particle density affects both systems. The results are significant for instances of good correlation between the particle density and velocity, such as occurs near the edge of a jet.
Integrated multiplexed capillary electrophoresis system
Yeung, Edward S.; Tan, Hongdong
2002-05-14
The present invention provides an integrated multiplexed capillary electrophoresis system for the analysis of sample analytes. The system integrates and automates multiple components, such as chromatographic columns and separation capillaries, and further provides a detector for the detection of analytes eluting from the separation capillaries. The system employs multiplexed freeze/thaw valves to manage fluid flow and sample movement. The system is computer controlled and is capable of processing samples through reaction, purification, denaturation, pre-concentration, injection, separation and detection in parallel fashion. Methods employing the system of the invention are also provided.
Elliott, Paul; Peakman, Tim C
2008-04-01
UK Biobank is a large prospective study in the UK to investigate the role of genetic factors, environmental exposures and lifestyle in the causes of major diseases of late and middle age. Extensive data and biological samples are being collected from 500,000 participants aged between 40 and 69 years. The biological samples that are collected and how they are processed and stored will have a major impact on the future scientific usefulness of the UK Biobank resource. The aim of the UK Biobank sample handling and storage protocol is to specify methods for the collection and storage of participant samples that give maximum scientific return within the available budget. Processing or storage methods that, as far as can be predicted, will preclude current or future assays have been avoided. The protocol was developed through a review of the literature on sample handling and processing, wide consultation within the academic community and peer review. Protocol development addressed which samples should be collected, how and when they should be processed and how the processed samples should be stored to ensure their long-term integrity. The recommended protocol was extensively tested in a series of validation studies. UK Biobank collects about 45 ml blood and 9 ml of urine with minimal local processing from each participant using the vacutainer system. A variety of preservatives, anti-coagulants and clot accelerators is used appropriate to the expected end use of the samples. Collection of other material (hair, nails, saliva and faeces) was also considered but rejected for the full cohort. Blood and urine samples from participants are transported overnight by commercial courier to a central laboratory where they are processed and aliquots of urine, plasma, serum, white cells and red cells stored in ultra-low temperature archives. Aliquots of whole blood are also stored for potential future production of immortalized cell lines. A standard panel of haematology assays is completed on whole blood from all participants, since such assays need to be conducted on fresh samples (whereas other assays can be done on stored samples). By the end of the recruitment phase, 15 million sample aliquots will be stored in two geographically separate archives: 9.5 million in a -80 degrees C automated archive and 5.5 million in a manual liquid nitrogen archive at -180 degrees C. Because of the size of the study and the numbers of samples obtained from participants, the protocol stipulates a highly automated approach for the processing and storage of samples. Implementation of the processes, technology, systems and facilities has followed best practices used in manufacturing industry to reduce project risk and to build in quality and robustness. The data produced from sample collection, processing and storage are highly complex and are managed by a commercially available LIMS system fully integrated with the entire process. The sample handling and storage protocol adopted by UK Biobank provides quality assured and validated methods that are feasible within the available funding and reflect the size and aims of the project. Experience from recruiting and processing the first 40,000 participants to the study demonstrates that the adopted methods and technologies are fit-for-purpose and robust.
NASA Astrophysics Data System (ADS)
Thurner, Stefan; Corominas-Murtra, Bernat; Hanel, Rudolf
2017-09-01
There are at least three distinct ways to conceptualize entropy: entropy as an extensive thermodynamic quantity of physical systems (Clausius, Boltzmann, Gibbs), entropy as a measure for information production of ergodic sources (Shannon), and entropy as a means for statistical inference on multinomial processes (Jaynes maximum entropy principle). Even though these notions represent fundamentally different concepts, the functional form of the entropy for thermodynamic systems in equilibrium, for ergodic sources in information theory, and for independent sampling processes in statistical systems, is degenerate, H (p ) =-∑ipilogpi . For many complex systems, which are typically history-dependent, nonergodic, and nonmultinomial, this is no longer the case. Here we show that for such processes, the three entropy concepts lead to different functional forms of entropy, which we will refer to as SEXT for extensive entropy, SIT for the source information rate in information theory, and SMEP for the entropy functional that appears in the so-called maximum entropy principle, which characterizes the most likely observable distribution functions of a system. We explicitly compute these three entropy functionals for three concrete examples: for Pólya urn processes, which are simple self-reinforcing processes, for sample-space-reducing (SSR) processes, which are simple history dependent processes that are associated with power-law statistics, and finally for multinomial mixture processes.
Assessing heat treatment of chicken breast cuts by impedance spectroscopy.
Schmidt, Franciny C; Fuentes, Ana; Masot, Rafael; Alcañiz, Miguel; Laurindo, João B; Barat, José M
2017-03-01
The aim of this work was to develop a new system based on impedance spectroscopy to assess the heat treatment of previously cooked chicken meat by two experiments; in the first, samples were cooked at different temperatures (from 60 to 90 ℃) until core temperature of the meat reached the water bath temperature. In the second approach, temperature was 80 ℃ and the samples were cooked for different times (from 5 to 55 min). Impedance was measured once samples had cooled. The examined processing parameters were the maximum temperature reached in thermal centre of the samples, weight loss, moisture and the integral of the temperature profile during the cooking-cooling process. The correlation between the processing parameters and impedance was studied by partial least square regressions. The models were able to predict the studied parameters. Our results are essential for developing a new system to control the technological, sensory and safety aspects of cooked meat products on the whole meat processing line.
Using the global positioning system to map disturbance patterns of forest harvesting machinery
T.P. McDonald; E.A. Carter; S.E. Taylor
2002-01-01
Abstract: A method was presented to transform sampled machine positional data obtained from a global positioning system (GPS) receiver into a two-dimensional raster map of number of passes as a function of location. The effect of three sources of error in the transformation process were investigated: path sampling rate (receiver sampling frequency);...
Process Research of Polycrystalline Silicon Material (PROPSM)
NASA Technical Reports Server (NTRS)
Culik, J. S.
1984-01-01
A passivation process (hydrogenation) that will improve the power generation of solar cells fabricated from presently produced, large grain, cast polycrystalline silicon (Semix), a potentially low cost material are developed. The first objective is to verify the operation of a DC plasma hydrogenation system and to investigate the effect of hydrogen on the electrical performance of a variety of polycrystalline silicon solar cells. The second objective is to parameterize and optimize a hydrogenation process for cast polycrystalline silicon, and will include a process sensitivity analysis. The sample preparation for the first phase is outlined. The hydrogenation system is described, and some early results that were obtained using the hydrogenation system without a plasma are summarized. Light beam induced current (LBIC) measurements of minicell samples, and their correlation to dark current voltage characteristics, are discussed.
The Society of Toxicologic Pathology charged a Nervous System Sampling Working Group with devising recommended practices to routinely screen the central and peripheral nervous systems in Good Laboratory Practice-type nonclinical general toxicity studies. Brains should be trimmed ...
40 CFR 65.113 - Standards: Sampling connection systems.
Code of Federal Regulations, 2011 CFR
2011-07-01
... be collected or captured. (c) Equipment design and operation. Each closed-purge, closed-loop, or... system; or (2) Collect and recycle the purged process fluid to a process; or (3) Be designed and operated to capture and transport all the purged process fluid to a control device that meets the requirements...
40 CFR 65.113 - Standards: Sampling connection systems.
Code of Federal Regulations, 2014 CFR
2014-07-01
... be collected or captured. (c) Equipment design and operation. Each closed-purge, closed-loop, or... system; or (2) Collect and recycle the purged process fluid to a process; or (3) Be designed and operated to capture and transport all the purged process fluid to a control device that meets the requirements...
40 CFR 65.113 - Standards: Sampling connection systems.
Code of Federal Regulations, 2010 CFR
2010-07-01
... be collected or captured. (c) Equipment design and operation. Each closed-purge, closed-loop, or... system; or (2) Collect and recycle the purged process fluid to a process; or (3) Be designed and operated to capture and transport all the purged process fluid to a control device that meets the requirements...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cornwell, Paris A; Bunn, Jeffrey R; Schmidlin, Joshua E
The December 2010 version of the guide, ORNL/TM-2008/159, by Jeff Bunn, Josh Schmidlin, Camden Hubbard, and Paris Cornwell, has been further revised due to a major change in the GeoMagic Studio software for constructing a surface model. The Studio software update also includes a plug-in module to operate the FARO Scan Arm. Other revisions for clarity were also made. The purpose of this revision document is to guide the reader through the process of laser alignment used by NRSF2 at HFIR and VULCAN at SNS. This system was created to increase the spatial accuracy of the measurement points in amore » sample, reduce the use of neutron time used for alignment, improve experiment planning, and reduce operator error. The need for spatial resolution has been driven by the reduction in gauge volumes to the sub-millimeter level, steep strain gradients in some samples, and requests to mount multiple samples within a few days for relating data from each sample to a common sample coordinate system. The first step in this process involves mounting the sample on an indexer table in a laboratory set up for offline sample mounting and alignment in the same manner it would be mounted at either instrument. In the shared laboratory, a FARO ScanArm is used to measure the coordinates of points on the sample surface ('point cloud'), specific features and fiducial points. A Sample Coordinate System (SCS) needs to be established first. This is an advantage of the technique because the SCS can be defined in such a way to facilitate simple definition of measurement points within the sample. Next, samples are typically mounted to a frame of 80/20 and fiducial points are attached to the sample or frame then measured in the established sample coordinate system. The laser scan probe on the ScanArm can then be used to scan in an 'as-is' model of the sample as well as mounting hardware. GeoMagic Studio 12 is the software package used to construct the model from the point cloud the scan arm creates. Once a model, fiducial, and measurement files are created, a special program, called SScanSS combines the information and by simulation of the sample on the diffractometer can help plan the experiment before using neutron time. Finally, the sample is mounted on the relevant stress measurement instrument and the fiducial points are measured again. In the HFIR beam room, a laser tracker is used in conjunction with a program called CAM2 to measure the fiducial points in the NRSF2 instrument's sample positioner coordinate system. SScanSS is then used again to perform a coordinate system transformation of the measurement file locations to the sample positioner coordinate system. A procedure file is then written with the coordinates in the sample positioner coordinate system for the desired measurement locations. This file is often called a script or command file and can be further modified using excel. It is very important to note that this process is not a linear one, but rather, it often is iterative. Many of the steps in this guide are interdependent on one another. It is very important to discuss the process as it pertains to the specific sample being measured. What works with one sample may not necessarily work for another. This guide attempts to provide a typical work flow that has been successful in most cases.« less
Using a computer controlled system, this ultrafiltration device automates the process of concentrating a water sample and can be operated in the field. The system was also designed to reduce human exposure to potentially contaminated water.
2011-12-01
systems engineering technical and technical management processes. Technical Planning, Stakeholders Requirements Development, and Architecture Design were...Stakeholder Requirements Definition, Architecture Design and Technical Planning. A purposive sampling of AFRL rapid development program managers and engineers...emphasize one process over another however Architecture Design , Implementation scored higher among Technical Processes. Decision Analysis, Technical
Note: A simple image processing based fiducial auto-alignment method for sample registration.
Robertson, Wesley D; Porto, Lucas R; Ip, Candice J X; Nantel, Megan K T; Tellkamp, Friedjof; Lu, Yinfei; Miller, R J Dwayne
2015-08-01
A simple method for the location and auto-alignment of sample fiducials for sample registration using widely available MATLAB/LabVIEW software is demonstrated. The method is robust, easily implemented, and applicable to a wide variety of experiment types for improved reproducibility and increased setup speed. The software uses image processing to locate and measure the diameter and center point of circular fiducials for distance self-calibration and iterative alignment and can be used with most imaging systems. The method is demonstrated to be fast and reliable in locating and aligning sample fiducials, provided here by a nanofabricated array, with accuracy within the optical resolution of the imaging system. The software was further demonstrated to register, load, and sample the dynamically wetted array.
NASA Astrophysics Data System (ADS)
Szydzik, C.; Gavela, A. F.; Roccisano, J.; Herranz de Andrés, S.; Mitchell, A.; Lechuga, L. M.
2016-12-01
We present recent results on the realisation and demonstration of an integrated optofluidic lab-on-a-chip measurement system. The system consists of an integrated on-chip automated microfluidic fluid handling subsystem, coupled with bimodal nano-interferometer waveguide technology, and is applied in the context of detection of antibiotics in seawater. The bimodal waveguide (BMWG) is a highly sensitive label-free biosensor. Integration of complex microfluidic systems with bimodal waveguide technology enables on-chip sample handling and fluid processing capabilities and allows for significant automation of experimental processes. The on-chip fluid-handling subsystem is realised through the integration of pneumatically actuated elastomer pumps and valves, enabling high temporal resolution sample and reagent delivery and facilitating multiplexed detection processes.
Shelton, Larry R.
1994-01-01
The U.S. Geological Survey's National Water-Quality Assessment program includes extensive data- collection efforts to assess the quality of the Nations's streams. These studies require analyses of stream samples for major ions, nutrients, sediments, and organic contaminants. For the information to be comparable among studies in different parts of the Nation, consistent procedures specifically designed to produce uncontaminated samples for trace analysis in the laboratory are critical. This field guide describes the standard procedures for collecting and processing samples for major ions, nutrients, organic contaminants, sediment, and field analyses of conductivity, pH, alkalinity, and dissolved oxygen. Samples are collected and processed using modified and newly designed equipment made of Teflon to avoid contamination, including nonmetallic samplers (D-77 and DH-81) and a Teflon sample splitter. Field solid-phase extraction procedures developed to process samples for organic constituent analyses produce an extracted sample with stabilized compounds for more accurate results. Improvements to standard operational procedures include the use of processing chambers and capsule filtering systems. A modified collecting and processing procedure for organic carbon is designed to avoid contamination from equipment cleaned with methanol. Quality assurance is maintained by strict collecting and processing procedures, replicate sampling, equipment blank samples, and a rigid cleaning procedure using detergent, hydrochloric acid, and methanol.
MSL's Widgets: Adding Rebustness to Martian Sample Acquisition, Handling, and Processing
NASA Technical Reports Server (NTRS)
Roumeliotis, Chris; Kennedy, Brett; Lin, Justin; DeGrosse, Patrick; Cady, Ian; Onufer, Nicholas; Sigel, Deborah; Jandura, Louise; Anderson, Robert; Katz, Ira;
2013-01-01
Mars Science Laboratory's (MSL) Sample Acquisition Sample Processing and Handling (SA-SPaH) system is one of the most ambitious terrain interaction and manipulation systems ever built and successfully used outside of planet earth. Mars has a ruthless environment that has surprised many who have tried to explore there. The robustness widget program was implemented by the MSL project to help ensure the SA-SPaH system would be robust enough to the surprises of this ruthless Martian environment. The robustness widget program was an effort of extreme schedule pressure and responsibility, but was accomplished with resounding success. This paper will focus on a behind the scenes look at MSL's robustness widgets: the particle fun zone, the wind guards, and the portioner pokers.
An Interactive Graphics Program for Investigating Digital Signal Processing.
ERIC Educational Resources Information Center
Miller, Billy K.; And Others
1983-01-01
Describes development of an interactive computer graphics program for use in teaching digital signal processing. The program allows students to interactively configure digital systems on a monitor display and observe their system's performance by means of digital plots on the system's outputs. A sample program run is included. (JN)
Oliver, C. Ryan; Westrick, William; Koehler, Jeremy; Brieland-Shoultz, Anna; Anagnostopoulos-Politis, Ilias; Cruz-Gonzalez, Tizoc; Hart, A. John
2013-01-01
Laboratory research and development on new materials, such as nanostructured thin films, often utilizes manual equipment such as tube furnaces due to its relatively low cost and ease of setup. However, these systems can be prone to inconsistent outcomes due to variations in standard operating procedures and limitations in performance such as heating and cooling rates restrict the parameter space that can be explored. Perhaps more importantly, maximization of research throughput and the successful and efficient translation of materials processing knowledge to production-scale systems, relies on the attainment of consistent outcomes. In response to this need, we present a semi-automated lab-scale chemical vapor deposition (CVD) furnace system, called “Robofurnace.” Robofurnace is an automated CVD system built around a standard tube furnace, which automates sample insertion and removal and uses motion of the furnace to achieve rapid heating and cooling. The system has a 10-sample magazine and motorized transfer arm, which isolates the samples from the lab atmosphere and enables highly repeatable placement of the sample within the tube. The system is designed to enable continuous operation of the CVD reactor, with asynchronous loading/unloading of samples. To demonstrate its performance, Robofurnace is used to develop a rapid CVD recipe for carbon nanotube (CNT) forest growth, achieving a 10-fold improvement in CNT forest mass density compared to a benchmark recipe using a manual tube furnace. In the long run, multiple systems like Robofurnace may be linked to share data among laboratories by methods such as Twitter. Our hope is Robofurnace and like automation will enable machine learning to optimize and discover relationships in complex material synthesis processes. PMID:24289435
Ion implantation system and process for ultrasensitive determination of target isotopes
DOE Office of Scientific and Technical Information (OSTI.GOV)
Farmer, III, Orville T.; Liezers, Martin
2016-09-13
A system and process are disclosed for ultrasensitive determination of target isotopes of analytical interest in a sample. Target isotopes may be implanted in an implant area on a high-purity substrate to pre-concentrate the target isotopes free of contaminants. A known quantity of a tracer isotope may also be implanted. Target isotopes and tracer isotopes may be determined in a mass spectrometer. The present invention provides ultrasensitive determination of target isotopes in the sample.
Optimizing the MAC Protocol in Localization Systems Based on IEEE 802.15.4 Networks
Claver, Jose M.; Ezpeleta, Santiago
2017-01-01
Radio frequency signals are commonly used in the development of indoor localization systems. The infrastructure of these systems includes some beacons placed at known positions that exchange radio packets with users to be located. When the system is implemented using wireless sensor networks, the wireless transceivers integrated in the network motes are usually based on the IEEE 802.15.4 standard. But, the CSMA-CA, which is the basis for the medium access protocols in this category of communication systems, is not suitable when several users want to exchange bursts of radio packets with the same beacon to acquire the radio signal strength indicator (RSSI) values needed in the location process. Therefore, new protocols are necessary to avoid the packet collisions that appear when multiple users try to communicate with the same beacons. On the other hand, the RSSI sampling process should be carried out very quickly because some systems cannot tolerate a large delay in the location process. This is even more important when the RSSI sampling process includes measures with different signal power levels or frequency channels. The principal objective of this work is to speed up the RSSI sampling process in indoor localization systems. To achieve this objective, the main contribution is the proposal of a new MAC protocol that eliminates the medium access contention periods and decreases the number of packet collisions to accelerate the RSSI collection process. Moreover, the protocol increases the overall network throughput taking advantage of the frequency channel diversity. The presented results show the suitability of this protocol for reducing the RSSI gathering delay and increasing the network throughput in simulated and real environments. PMID:28684666
Optimizing the MAC Protocol in Localization Systems Based on IEEE 802.15.4 Networks.
Pérez-Solano, Juan J; Claver, Jose M; Ezpeleta, Santiago
2017-07-06
Radio frequency signals are commonly used in the development of indoor localization systems. The infrastructure of these systems includes some beacons placed at known positions that exchange radio packets with users to be located. When the system is implemented using wireless sensor networks, the wireless transceivers integrated in the network motes are usually based on the IEEE 802.15.4 standard. But, the CSMA-CA, which is the basis for the medium access protocols in this category of communication systems, is not suitable when several users want to exchange bursts of radio packets with the same beacon to acquire the radio signal strength indicator (RSSI) values needed in the location process. Therefore, new protocols are necessary to avoid the packet collisions that appear when multiple users try to communicate with the same beacons. On the other hand, the RSSI sampling process should be carried out very quickly because some systems cannot tolerate a large delay in the location process. This is even more important when the RSSI sampling process includes measures with different signal power levels or frequency channels. The principal objective of this work is to speed up the RSSI sampling process in indoor localization systems. To achieve this objective, the main contribution is the proposal of a new MAC protocol that eliminates the medium access contention periods and decreases the number of packet collisions to accelerate the RSSI collection process. Moreover, the protocol increases the overall network throughput taking advantage of the frequency channel diversity. The presented results show the suitability of this protocol for reducing the RSSI gathering delay and increasing the network throughput in simulated and real environments.
ROMPS critical design review data package
NASA Technical Reports Server (NTRS)
Dobbs, M. E.
1992-01-01
The design elements of the Robot-Operated Material Processing in Space (ROMPS) system are described in outline and graphical form. The following subsystems/topics are addressed: servo system, testbed and simulation results, System V Controller, robot module, furnace module, SCL experiment supervisor and script sample processing control, battery system, watchdog timers, mechanical/thermal considerations, and fault conditions and recovery.
40 CFR 63.1013 - Sampling connection systems standards.
Code of Federal Regulations, 2011 CFR
2011-07-01
... (CONTINUED) National Emission Standards for Equipment Leaks-Control Level 1 § 63.1013 Sampling connection... container are not required to be collected or captured. (c) Equipment design and operation. Each closed... process fluid to a process; or (3) Be designed and operated to capture and transport all the purged...
Systems and methods for laser assisted sample transfer to solution for chemical analysis
Van Berkel, Gary J.; Kertesz, Vilmos; Ovchinnikova, Olga S.
2014-06-03
Systems and methods are described for laser ablation of an analyte from a specimen and capturing of the analyte in a dispensed solvent to form a testing solution. A solvent dispensing and extraction system can form a liquid microjunction with the specimen. The solvent dispensing and extraction system can include a surface sampling probe. The laser beam can be directed through the surface sampling probe. The surface sampling probe can also serve as an atomic force microscopy probe. The surface sampling probe can form a seal with the specimen. The testing solution including the analyte can then be analyzed using an analytical instrument or undergo further processing.
Systems and methods for laser assisted sample transfer to solution for chemical analysis
Van Berkel, Gary J.; Kertesz, Vilmos; Ovchinnikova, Olga S.
2015-09-29
Systems and methods are described for laser ablation of an analyte from a specimen and capturing of the analyte in a dispensed solvent to form a testing solution. A solvent dispensing and extraction system can form a liquid microjunction with the specimen. The solvent dispensing and extraction system can include a surface sampling probe. The laser beam can be directed through the surface sampling probe. The surface sampling probe can also serve as an atomic force microscopy probe. The surface sampling probe can form a seal with the specimen. The testing solution including the analyte can then be analyzed using an analytical instrument or undergo further processing.
Systems and methods for laser assisted sample transfer to solution for chemical analysis
Van Berkel, Gary J; Kertesz, Vilmos; Ovchinnikova, Olga S
2013-08-27
Systems and methods are described for laser ablation of an analyte from a specimen and capturing of the analyte in a dispensed solvent to form a testing solution. A solvent dispensing and extraction system can form a liquid microjunction with the specimen. The solvent dispensing and extraction system can include a surface sampling probe. The laser beam can be directed through the surface sampling probe. The surface sampling probe can also serve as an atomic force microscopy probe. The surface sampling probe can form a seal with the specimen. The testing solution including the analyte can then be analyzed using an analytical instrument or undergo further processing.
What is Next? Linking all Samples of Planet Earth.
NASA Astrophysics Data System (ADS)
Wyborn, L. A.; Lehnert, K.; Klump, J. F.; Arko, R. A.; Cox, S. J. D.; Devaraju, A.; Elger, K.; Murphy, F.; Fleischer, D.
2016-12-01
The process of sampling, observing and analyzing physical samples is not unique to the geosciences. Physical sampling (taking specimens) is a fundamental strategy in many natural sciences, typically to support ex-situ observations in laboratories with the goal of characterizing real-world entities or populations. Observations and measurements are made on individual specimens and their derived samples in various ways, with results reported in research publications. Research on an individual sample is often published in numerous articles, based on multiple, potentially unrelated research programs conducted over many years. Even high-volume Earth observation datasets are proxies of real world phenomena and require calibration by measurements made on position located, well described physical samples. Unique, persistent web-compatible identifiers for physical objects and related sampling features are required to ensure their unambiguous citation and connection to related datasets through web identifiers. Identifier systems have been established within specific domains (e.g., bio, geo, hydro) or different sectors (e.g., museums, government agencies, universities), including the International Geo Sample Number (IGSN) in the geosciences, which has been used for rock, fossil, mineral, soil, regolith, fluid, plant and synthetic materials. IGSNs are issued through a governance system that ensures they are globally unique. Each IGSN directs to a digital representation of the physical object via the Handle.net global resolver system, the same system used for resolving DOI. To enable the unique identification of all samples on Planet Earth and of data derived from them, the next step is to ensure IGSNs can either be integrated with comparable identifier systems in other domains/sectors, or introduced into domains that do not have a viable system. A registry of persistent identifier systems for physical samples would allow users to choose which system best suits their needs. Such a registry may also facilitate unifying best practice in these multiple systems to enable consistent referencing of physical samples and of methods used to link digital data to its sources. IGSNs could be extended into other domains, but additional methodologies of sample collection, curation and processing may need to be considered.
Potato Operation: automatic detection of potato diseases
NASA Astrophysics Data System (ADS)
Lefebvre, Marc; Zimmerman, Thierry; Baur, Charles; Guegerli, Paul; Pun, Thierry
1995-01-01
The Potato Operation is a collaborative, multidisciplinary project in the domain of destructive testing of agricultural products. It aims at automatizing pulp sampling of potatoes in order to detect possible viral diseases. Such viruses can decrease fields productivity by a factor of up to ten. A machine, composed of three conveyor belts, a vision system, a robotic arm and controlled by a PC has been built. Potatoes are brought one by one from a bulk to the vision system, where they are seized by a rotating holding device. The sprouts, where the viral activity is maximum, are then detected by an active vision process operating on multiple views. The 3D coordinates of the sampling point are communicated to the robot arm holding a drill. Some flesh is then sampled by the drill, then deposited into an Elisa plate. After sampling, the robot arm washes the drill in order to prevent any contamination. The PC computer simultaneously controls these processes, the conveying of the potatoes, the vision algorithms and the sampling procedure. The master process, that is the vision procedure, makes use of three methods to achieve the sprouts detection. A profile analysis first locates the sprouts as protuberances. Two frontal analyses, respectively based on fluorescence and local variance, confirm the previous detection and provide the 3D coordinate of the sampling zone. The other two processes work by interruption of the master process.
Chango, Gabriela; Palacio, Edwin; Cerdà, Víctor
2018-08-15
A simple potentiometric chip-based multipumping flow system (MPFS) has been developed for the simultaneous determination of fluoride, chloride, pH, and redox potential in water samples. The proposed system was developed by using a poly(methyl methacrylate) chip microfluidic-conductor using the advantages of flow techniques with potentiometric detection. For this purpose, an automatic system has been designed and built by optimizing the variables involved in the process, such as: pH, ionic strength, stirring and sample volume. This system was applied successfully to water samples getting a versatile system with an analysis frequency of 12 samples per hour. Good correlation between chloride and fluoride concentration measured with ISE and ionic chromatography technique suggests satisfactory reliability of the system. Copyright © 2018 Elsevier B.V. All rights reserved.
Continuous-flow free acid monitoring method and system
Strain, J.E.; Ross, H.H.
1980-01-11
A free acid monitoring method and apparatus is provided for continuously measuring the excess acid present in a process stream. The disclosed monitoring system and method is based on the relationship of the partial pressure ratio of water and acid in equilibrium with an acid solution at constant temperature. A portion of the process stream is pumped into and flows through the monitor under the influence of gravity and back to the process stream. A continuous flowing sample is vaporized at a constant temperature and the vapor is subsequently condensed. Conductivity measurements of the condensate produces a nonlinear response function from which the free acid molarity of the sample process stream is determined.
Continuous-flow free acid monitoring method and system
Strain, James E.; Ross, Harley H.
1981-01-01
A free acid monitoring method and apparatus is provided for continuously measuring the excess acid present in a process stream. The disclosed monitoring system and method is based on the relationship of the partial pressure ratio of water and acid in equilibrium with an acid solution at constant temperature. A portion of the process stream is pumped into and flows through the monitor under the influence of gravity and back to the process stream. A continuous flowing sample is vaporized at a constant temperature and the vapor is subsequently condensed. Conductivity measurements of the condensate produces a nonlinear response function from which the free acid molarity of the sample process stream is determined.
DOE Office of Scientific and Technical Information (OSTI.GOV)
TERRI, FELLINGER
2004-12-21
The Defense Waste Processing Facility, DWPF, currently generates approximately 1.4 million gallons of recycle water per year during Sludge-Only operations. DWPF has minimized condensate generation to 1.4 million gallons by not operating the Steam Atomized Scrubbers, SASs, for the melter off gas system. By not operating the SASs, DWPF has reduced the total volume by approximately 800,000 gallons of condensate per year. Currently, the recycle stream is sent to back to the Tank Farm and processed through the 2H Evaporator system. To alleviate the load on the 2H Evaporator system, an acid evaporator design is being considered as an alternatemore » processing and/or concentration method for the DWPF recycle stream. In order to support this alternate processing option, the DWPF has requested that the chemical and radionuclide compositions of the Off Gas Condensate Tank, OGCT, Slurry Mix Evaporator Condensate Tank, SMECT, Recycle Collection Tank, RCT, and the Decontamination Waste Treatment Tank, DWTT, be determined as a part of the process development work for the acid evaporator design. Samples have been retrieved from the OGCT, RCT, and SMECT and have been sent to the Savannah River National Laboratory, SRNL for this characterization. The DWTT samples have been recently shipped to SRNL. The results for the DWTT samples will be issued at later date.« less
Advanced Engine Health Management Applications of the SSME Real-Time Vibration Monitoring System
NASA Technical Reports Server (NTRS)
Fiorucci, Tony R.; Lakin, David R., II; Reynolds, Tracy D.; Turner, James E. (Technical Monitor)
2000-01-01
The Real Time Vibration Monitoring System (RTVMS) is a 32-channel high speed vibration data acquisition and processing system developed at Marshall Space Flight Center (MSFC). It Delivers sample rates as high as 51,200 samples/second per channel and performs Fast Fourier Transform (FFT) processing via on-board digital signal processing (DSP) chips in a real-time format. Advanced engine health assessment is achieved by utilizing the vibration spectra to provide accurate sensor validation and enhanced engine vibration redlines. Discrete spectral signatures (such as synchronous) that are indicators of imminent failure can be assessed and utilized to mitigate catastrophic engine failures- a first in rocket engine health assessment. This paper is presented in viewgraph form.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oden, L.L.; O`Conner, W.K.; Turner, P.C.
1993-11-19
This report presents field results and raw data from the Buried Waste Integrated Demonstration (BWID) Arc Melter Vitrification Project Phase 1 baseline test series conducted by the Idaho National Engineering Laboratory (INEL) in cooperation with the U.S. Bureau of Mines (USBM). The baseline test series was conducted using the electric arc melter facility at the USBM Albany Research Center in Albany, Oregon. Five different surrogate waste feed mixtures were tested that simulated thermally-oxidized, buried, TRU-contaminated, mixed wastes and soils present at the INEL. The USBM Arc Furnace Integrated Waste Processing Test Facility includes a continuous feed system, the arc meltingmore » furnace, an offgas control system, and utilities. The melter is a sealed, 3-phase alternating current (ac) furnace approximately 2 m high and 1.3 m wide. The furnace has a capacity of 1 metric ton of steel and can process as much as 1,500 lb/h of soil-type waste materials. The surrogate feed materials included five mixtures designed to simulate incinerated TRU-contaminated buried waste materials mixed with INEL soil. Process samples, melter system operations data and offgas composition data were obtained during the baseline tests to evaluate the melter performance and meet test objectives. Samples and data gathered during this program included (a) automatically and manually logged melter systems operations data, (b) process samples of slag, metal and fume solids, and (c) offgas composition, temperature, velocity, flowrate, moisture content, particulate loading and metals content. This report consists of 2 volumes: Volume I summarizes the baseline test operations. It includes an executive summary, system and facility description, review of the surrogate waste mixtures, and a description of the baseline test activities, measurements, and sample collection. Volume II contains the raw test data and sample analyses from samples collected during the baseline tests.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oden, L.L.; O`Connor, W.K.; Turner, P.C.
1993-11-19
This report presents field results and raw data from the Buried Waste Integrated Demonstration (BWID) Arc Melter Vitrification Project Phase 1 baseline test series conducted by the Idaho National Engineering Laboratory (INEL) in cooperation with the U.S. Bureau of Mines (USBM). The baseline test series was conducted using the electric arc melter facility at the USBM Albany Research Center in Albany, Oregon. Five different surrogate waste feed mixtures were tested that simulated thermally-oxidized, buried, TRU-contaminated, mixed wastes and soils present at the INEL. The USBM Arc Furnace Integrated Waste Processing Test Facility includes a continuous feed system, the arc meltingmore » furnace, an offgas control system, and utilities. The melter is a sealed, 3-phase alternating current (ac) furnace approximately 2 m high and 1.3 m wide. The furnace has a capacity of 1 metric ton of steel and can process as much as 1,500 lb/h of soil-type waste materials. The surrogate feed materials included five mixtures designed to simulate incinerated TRU-contaminated buried waste materials mixed with INEL soil. Process samples, melter system operations data and offgas composition data were obtained during the baseline tests to evaluate the melter performance and meet test objectives. Samples and data gathered during this program included (a) automatically and manually logged melter systems operations data, (b) process samples of slag, metal and fume solids, and (c) offgas composition, temperature, velocity, flowrate, moisture content, particulate loading and metals content. This report consists of 2 volumes: Volume I summarizes the baseline test operations. It includes an executive summary, system and facility description, review of the surrogate waste mixtures, and a description of the baseline test activities, measurements, and sample collection. Volume II contains the raw test data and sample analyses from samples collected during the baseline tests.« less
Automated soil gas monitoring chamber
Edwards, Nelson T.; Riggs, Jeffery S.
2003-07-29
A chamber for trapping soil gases as they evolve from the soil without disturbance to the soil and to the natural microclimate within the chamber has been invented. The chamber opens between measurements and therefore does not alter the metabolic processes that influence soil gas efflux rates. A multiple chamber system provides for repetitive multi-point sampling, undisturbed metabolic soil processes between sampling, and an essentially airtight sampling chamber operating at ambient pressure.
NASA Astrophysics Data System (ADS)
Jia, Bing; Wei, Jian-Ping; Wen, Zhi-Hui; Wang, Yun-Gang; Jia, Lin-Xing
2017-11-01
In order to study the response characteristics of infrasound in coal samples under the uniaxial loading process, coal samples were collected from GengCun mine. Coal rock stress loading device, acoustic emission tested system and infrasound tested system were used to test the infrasonic signal and acoustic emission signal under uniaxial loading process. The tested results were analyzed by the methods of wavelet filter, threshold denoise, time-frequency analysis and so on. The results showed that in the loading process, the change of the infrasonic wave displayed the characteristics of stage, and it could be divided into three stages: initial stage with a certain amount infrasound events, middle stage with few infrasound events, and late stage gradual decrease. It had a good consistency with changing characteristics of acoustic emission. At the same time, the frequency of infrasound was very low. It can propagate over a very long distance with little attenuation, and the characteristics of the infrasound before the destruction of the coal samples were obvious. A method of using the infrasound characteristics to predict the destruction of coal samples was proposed. This is of great significance to guide the prediction of geological hazards in coal mines.
System and process for dissolution of solids
Liezers, Martin; Farmer, III, Orville T.
2017-10-10
A system and process are disclosed for dissolution of solids and "difficult-to-dissolve" solids. A solid sample may be ablated in an ablation device to generate nanoscale particles. Nanoparticles may then swept into a coupled plasma device operating at atmospheric pressure where the solid nanoparticles are atomized. The plasma exhaust may be delivered directly into an aqueous fluid to form a solution containing the atomized and dissolved solids. The composition of the resulting solution reflects the composition of the original solid sample.
An automated atmospheric sampling system operating on 747 airliners
NASA Technical Reports Server (NTRS)
Perkins, P. J.; Gustafsson, U. R. C.
1976-01-01
An air sampling system that automatically measures the temporal and spatial distribution of particulate and gaseous constituents of the atmosphere is collecting data on commercial air routes covering the world. Measurements are made in the upper troposphere and lower stratosphere (6 to 12 km) of constituents related to aircraft engine emissions and other pollutants. Aircraft operated by different airlines sample air at latitudes from the Arctic to Australia. This unique system includes specialized instrumentation, a special air inlet probe for sampling outside air, a computerized automatic control, and a data acquisition system. Air constituent and related flight data are tape recorded in flight for later computer processing on the ground.
Immune system changes during simulated planetary exploration on Devon Island, high arctic
Crucian, Brian; Lee, Pascal; Stowe, Raymond; Jones, Jeff; Effenhauser, Rainer; Widen, Raymond; Sams, Clarence
2007-01-01
Background Dysregulation of the immune system has been shown to occur during spaceflight, although the detailed nature of the phenomenon and the clinical risks for exploration class missions have yet to be established. Also, the growing clinical significance of immune system evaluation combined with epidemic infectious disease rates in third world countries provides a strong rationale for the development of field-compatible clinical immunology techniques and equipment. In July 2002 NASA performed a comprehensive immune assessment on field team members participating in the Haughton-Mars Project (HMP) on Devon Island in the high Canadian Arctic. The purpose of the study was to evaluate the effect of mission-associated stressors on the human immune system. To perform the study, the development of techniques for processing immune samples in remote field locations was required. Ten HMP-2002 participants volunteered for the study. A field protocol was developed at NASA-JSC for performing sample collection, blood staining/processing for immunophenotype analysis, whole-blood mitogenic culture for functional assessments and cell-sample preservation on-location at Devon Island. Specific assays included peripheral leukocyte distribution; constitutively activated T cells, intracellular cytokine profiles, plasma cortisol and EBV viral antibody levels. Study timepoints were 30 days prior to mission start, mid-mission and 60 days after mission completion. Results The protocol developed for immune sample processing in remote field locations functioned properly. Samples were processed on Devon Island, and stabilized for subsequent analysis at the Johnson Space Center in Houston. The data indicated that some phenotype, immune function and stress hormone changes occurred in the HMP field participants that were largely distinct from pre-mission baseline and post-mission recovery data. These immune changes appear similar to those observed in astronauts following spaceflight. Conclusion The immune system changes described during the HMP field deployment validate the use of the HMP as a ground-based spaceflight/planetary exploration analog for some aspects of human physiology. The sample processing protocol developed for this study may have applications for immune studies in remote terrestrial field locations. Elements of this protocol could possibly be adapted for future in-flight immunology studies conducted during space missions. PMID:17521440
System and method for measuring fluorescence of a sample
Riot, Vincent J
2015-03-24
The present disclosure provides a system and a method for measuring fluorescence of a sample. The sample may be a polymerase-chain-reaction (PCR) array, a loop-mediated-isothermal amplification array, etc. LEDs are used to excite the sample, and a photodiode is used to collect the sample's fluorescence. An electronic offset signal is used to reduce the effects of background fluorescence and the noises from the measurement system. An integrator integrates the difference between the output of the photodiode and the electronic offset signal over a given period of time. The resulting integral is then converted into digital domain for further processing and storage.
40 CFR 86.537-90 - Dynamometer test runs.
Code of Federal Regulations, 2014 CFR
2014-07-01
... “transient” formaldehyde exhaust sample, the “transient” dilution air sample bag, the “transient” methanol... start “transient” exhaust and dilution air bag samples to the analytical system and process the samples... Section 86.537-90 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS...
40 CFR 86.537-90 - Dynamometer test runs.
Code of Federal Regulations, 2012 CFR
2012-07-01
... “transient” formaldehyde exhaust sample, the “transient” dilution air sample bag, the “transient” methanol... start “transient” exhaust and dilution air bag samples to the analytical system and process the samples... Section 86.537-90 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS...
40 CFR 86.537-90 - Dynamometer test runs.
Code of Federal Regulations, 2013 CFR
2013-07-01
... “transient” formaldehyde exhaust sample, the “transient” dilution air sample bag, the “transient” methanol... start “transient” exhaust and dilution air bag samples to the analytical system and process the samples... Section 86.537-90 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS...
Immiscible phase incorporation during directional solidification of hypermonotectics
NASA Technical Reports Server (NTRS)
Andrews, J. Barry; Merrick, Roger A.
1993-01-01
Solidification processes in immiscible samples were investigated by directly observing the events taking place at the solid-liquid interface during directional solidification. Visualization of these events was made possible through the use of a transparent metal analog system and a temperature gradient stage assembly fitted to an optical microscope. The immiscible transparent analog system utilized was the succinonitrile-glycerol system. This system has been shown to exhibit the same morphological transitions as observed in metallic alloys of monotectic composition. Both monotectic and hypermonotectic composition samples were directionally solidified in order to gain an improved understanding of the manner in which the excess hypermonotectic liquid is incorporated into the solidifying structure. The processing conditions utilized prevented sedimentation of the excess hypermonotectic liquid by directionally solidifying the samples in very thin (13 microns), horizontally oriented cells. High thermal gradient to growth rate ratios (G/R) were used in an effort to prevent constitutional supercooling and the subsequent formation of L(sub 2) droplets in advance of the solidification front during the growth of fibrous composite structures. Results demonstrated that hypermonotectic composites could be produced in samples up to two weight percent off of the monotectic composition by using a G/R ratio greater than or equal to 4.6 x 10(exp 4) C(s)/mm(sup 2) to avoid constitutional supercooling. For hypermonotectic samples processed with G/R ratios below 4.6 x 10(exp 4) C(s)/mm(sup 2), constitutional supercooling occurred and resulted in slight interfacial instability. For these samples, two methods of incorporation of the hypermonotectic liquid were observed and are reported. The correlation between the phase spacing, lambda, and the growth rate, R, was examined and was found to obey a relationship generally associated with a diffusion controlled coupled growth process. For samples with compositions ranging from the monotectic composition up to 2 percent off of the monotectic composition, data indicated that the square of the phase spacing (lambda) varied linearly with the inverse of the growth rate (R).
NASA Astrophysics Data System (ADS)
Lee, T. J.; Lee, K. S., , Dr; Lee, S. K.
2017-12-01
One of the most important factors in measuring effective porosity by vacuum saturation method is that the air in the pore space can be fully substituted by water during the vacuum saturation process. International Society of Rock Mechanics (ISRM) suggests vacuuming a rock sample submerged in the water, while American Society of Test and Materials (ASTM) vacuuming the sample and water separately and then pour the water to the sample. In this study, we call the former wet-type vacuum saturation (WVS) method and the latter dry-type vacuum saturation (DVS) method, and compare the effective porosity measured by the two different vacuum saturation processes. For that purpose, a vacuum saturation system has been developed, which can support both WVS and DVS by only changing the process by programming. Comparison of effective porosity has been made for a cement mortar and rock samples. As a result, DVS can substitute more void volume to water than WVS, which in turn insists that DVS can provide more exact value of effective porosity than WVS.
NASA Astrophysics Data System (ADS)
Josephsen, Gary D.; Josephsen, Kelly A.; Beilman, Greg J.; Taylor, Jodie H.; Muiler, Kristine E.
2005-12-01
This is a report of the adaptation of microwave processing in the preparation of liver biopsies for transmission electron microscopy (TEM) to examine ultrastructural damage of mitochondria in the setting of metabolic stress. Hemorrhagic shock was induced in pigs via 35% total blood volume bleed and a 90-min period of shock followed by resuscitation. Hepatic biopsies were collected before shock and after resuscitation. Following collection, biopsies were processed for TEM by a rapid method involving microwave irradiation (Giberson, 2001). Samples pre- and postshock of each of two animals were viewed and scored using the mitochondrial ultrastructure scoring system (Crouser et al., 2002), a system used to quantify the severity of ultrastructural damage during shock. Results showed evidence of increased ultrastructural damage in the postshock samples, which scored 4.00 and 3.42, versus their preshock controls, which scored 1.18 and 1.27. The results of this analysis were similar to those obtained in another model of shock (Crouser et al., 2002). However, the amount of time used to process the samples was significantly shortened with methods involving microwave irradiation.
NASA Astrophysics Data System (ADS)
Nelson, E.; L'Ecuyer, T. S.; Wood, N.; Smalley, M.; Kulie, M.; Hahn, W.
2017-12-01
Global models exhibit substantial biases in the frequency, intensity, duration, and spatial scales of precipitation systems. Much of this uncertainty stems from an inadequate representation of the processes by which water is cycled between the surface and atmosphere and, in particular, those that govern the formation and maintenance of cloud systems and their propensity to form the precipitation. Progress toward improving precipitation process models requires observing systems capable of quantifying the coupling between the ice content, vertical mass fluxes, and precipitation yield of precipitating cloud systems. Spaceborne multi-frequency, Doppler radar offers a unique opportunity to address this need but the effectiveness of such a mission is heavily dependent on its ability to actually observe the processes of interest in the widest possible range of systems. Planning for a next generation precipitation process observing system should, therefore, start with a fundamental evaluation of the trade-offs between sensitivity, resolution, sampling, cost, and the overall potential scientific yield of the mission. Here we provide an initial assessment of the scientific and economic trade-space by evaluating hypothetical spaceborne multi-frequency radars using a combination of current real-world and model-derived synthetic observations. Specifically, we alter the field of view, vertical resolution, and sensitivity of a hypothetical Ka- and W-band radar system and propagate those changes through precipitation detection and intensity retrievals. The results suggest that sampling biases introduced by reducing sensitivity disproportionately affect the light rainfall and frozen precipitation regimes that are critical for warm cloud feedbacks and ice sheet mass balance, respectively. Coarser spatial resolution observations introduce regime-dependent biases in both precipitation occurrence and intensity that depend on cloud regime, with even the sign of the bias varying within a single storm system. It is suggested that the next generation spaceborne radar have a minimum sensitivity of -5 dBZ and spatial resolution of at least 3 km at all frequencies to adequately sample liquid and ice phase precipitation processes globally.
On incomplete sampling under birth-death models and connections to the sampling-based coalescent.
Stadler, Tanja
2009-11-07
The constant rate birth-death process is used as a stochastic model for many biological systems, for example phylogenies or disease transmission. As the biological data are usually not fully available, it is crucial to understand the effect of incomplete sampling. In this paper, we analyze the constant rate birth-death process with incomplete sampling. We derive the density of the bifurcation events for trees on n leaves which evolved under this birth-death-sampling process. This density is used for calculating prior distributions in Bayesian inference programs and for efficiently simulating trees. We show that the birth-death-sampling process can be interpreted as a birth-death process with reduced rates and complete sampling. This shows that joint inference of birth rate, death rate and sampling probability is not possible. The birth-death-sampling process is compared to the sampling-based population genetics model, the coalescent. It is shown that despite many similarities between these two models, the distribution of bifurcation times remains different even in the case of very large population sizes. We illustrate these findings on an Hepatitis C virus dataset from Egypt. We show that the transmission times estimates are significantly different-the widely used Gamma statistic even changes its sign from negative to positive when switching from the coalescent to the birth-death process.
40 CFR 63.104 - Heat exchange system requirements.
Code of Federal Regulations, 2012 CFR
2012-07-01
... heat exchange system or at locations where the cooling water enters and exits each heat exchanger or any combination of heat exchangers. (i) For samples taken at the entrance and exit of recirculating... manufacturing process units. (iii) For samples taken at the entrance and exit of each heat exchanger or any...
Fixture For Sampling Volatile Materials In Containers
NASA Technical Reports Server (NTRS)
Melton, Donald; Pratz, Earl Howard
1995-01-01
Fixture based on T-connector enables mass-spectrometric analysis of volatile contents of cylindrical containers without exposing contents to ambient conditions. Used to sample volatile contents of pressurized containers, contents of such enclosed processing systems as gas-phase reactors, gases in automotive emission systems, and gas in hostile environments.
40 CFR 63.1032 - Sampling connection systems standards.
Code of Federal Regulations, 2013 CFR
2013-07-01
...) [Reserved] (3) Be designed and operated to capture and transport all the purged process fluid to a control... (CONTINUED) National Emission Standards for Equipment Leaks-Control Level 2 Standards § 63.1032 Sampling... design and operation. Each closed-purge, closed-loop, or closed vent system as required in paragraph (b...
40 CFR 63.1032 - Sampling connection systems standards.
Code of Federal Regulations, 2011 CFR
2011-07-01
...) [Reserved] (3) Be designed and operated to capture and transport all the purged process fluid to a control... (CONTINUED) National Emission Standards for Equipment Leaks-Control Level 2 Standards § 63.1032 Sampling... design and operation. Each closed-purge, closed-loop, or closed vent system as required in paragraph (b...
Diffraction patterns in Fresnel approximation of periodic objects for a colorimeter of two apertures
NASA Astrophysics Data System (ADS)
Cortes-Reynoso, Jose-German R.; Suarez-Romero, Jose G.; Hurtado-Ramos, Juan B.; Tepichin-Rodriguez, Eduardo; Solorio-Leyva, Juan Carlos
2004-10-01
In this work, we present a study of Fresnel diffraction of periodic structures in an optical system of two apertures. This system of two apertures was used successfully for measuring color in textile samples solving the problems of illumination and directionality that present current commercial equipments. However, the system is sensible to the spatial frequency of the periodic sample"s area enclosed in its optical field of view. The study of Fresnel diffraction allows us to establish criteria for geometrical parameters of measurements in order to assure invariance in angular rotations and spatial positions. In this work, we use the theory of partial coherence to calculate the diffraction through two continuous apertures. In the calculation process, we use the concept of point-spread function of the system for partial coherence, in this way we avoid complicated statistical processes commonly used in the partial coherence theory.
In-flight edge response measurements for high-spatial-resolution remote sensing systems
NASA Astrophysics Data System (ADS)
Blonski, Slawomir; Pagnutti, Mary A.; Ryan, Robert; Zanoni, Vickie
2002-09-01
In-flight measurements of spatial resolution were conducted as part of the NASA Scientific Data Purchase Verification and Validation process. Characterization included remote sensing image products with ground sample distance of 1 meter or less, such as those acquired with the panchromatic imager onboard the IKONOS satellite and the airborne ADAR System 5500 multispectral instrument. Final image products were used to evaluate the effects of both the image acquisition system and image post-processing. Spatial resolution was characterized by full width at half maximum of an edge-response-derived line spread function. The edge responses were analyzed using the tilted-edge technique that overcomes the spatial sampling limitations of the digital imaging systems. As an enhancement to existing algorithms, the slope of the edge response and the orientation of the edge target were determined by a single computational process. Adjacent black and white square panels, either painted on a flat surface or deployed as tarps, formed the ground-based edge targets used in the tests. Orientation of the deployable tarps was optimized beforehand, based on simulations of the imaging system. The effects of such factors as acquisition geometry, temporal variability, Modulation Transfer Function compensation, and ground sample distance on spatial resolution were investigated.
Self-Sealing Wet Chemistry Cell for Field Analysis
NASA Technical Reports Server (NTRS)
Beegle, Luther W.; Soto, Juancarlos; Lasnik, James; Roark, Shane
2012-01-01
In most analytical investigations, there is a need to process complex field samples for the unique detection of analytes, especially when detecting low concentration organic molecules that may identify extraterrestrial life. Wet chemistry based instruments are the techniques of choice for most laboratory- based analysis of organic molecules due to several factors including less fragmentation of fragile biomarkers, and ability to concentrate target species resulting in much lower limits of detection. Development of an automated wet chemistry preparation system that can operate autonomously on Earth and is also designed to operate under Martian ambient conditions will demonstrate the technical feasibility of including wet chemistry on future missions. An Automated Sample Processing System (ASPS) has recently been developed that receives fines, extracts organics through solvent extraction, processes the extract by removing non-organic soluble species, and delivers sample to multiple instruments for analysis (including for non-organic soluble species). The key to this system is a sample cell that can autonomously function under field conditions. As a result, a self-sealing sample cell was developed that can autonomously hermetically seal fines and powder into a container, regardless of orientation of the apparatus. The cap is designed with a beveled edge, which allows the cap to be self-righted as the capping motor engages. Each cap consists of a C-clip lock ring below a crucible O-ring that is placed into a groove cut into the sample cap.
A Combined Fabrication and Instrumentation Platform for Sample Preparation.
Guckenberger, David J; Thomas, Peter C; Rothbauer, Jacob; LaVanway, Alex J; Anderson, Meghan; Gilson, Dan; Fawcett, Kevin; Berto, Tristan; Barrett, Kevin; Beebe, David J; Berry, Scott M
2014-06-01
While potentially powerful, access to molecular diagnostics is substantially limited in the developing world. Here we present an approach to reduced cost molecular diagnostic instrumentation that has the potential to empower developing world communities by reducing costs through streamlining the sample preparation process. In addition, this instrument is capable of producing its own consumable devices on demand, reducing reliance on assay suppliers. Furthermore, this instrument is designed with an "open" architecture, allowing users to visually observe the assay process and make modifications as necessary (as opposed to traditional "black box" systems). This open environment enables integration of microfluidic fabrication and viral RNA purification onto an easy-to-use modular system via the use of interchangeable trays. Here we employ this system to develop a protocol to fabricate microfluidic devices and then use these devices to isolate viral RNA from serum for the measurement of human immunodeficiency virus (HIV) viral load. Results obtained from this method show significantly reduced error compared with similar nonautomated sample preparation processes. © 2014 Society for Laboratory Automation and Screening.
40 CFR Appendix B to Part 63 - Sources Defined for Early Reduction Provisions
Code of Federal Regulations, 2010 CFR
2010-07-01
.... All valves in gas or light liquid service within a process unit b. All pumps in light liquid service within a process unit c. All connectors in gas or light liquid service within a process unit d. Each...-ended valve or line i. Each sampling connection system j. Each instrumentation system k. Each pump...
Davis, Ben; Grosvenor, Chriss; Johnk, Robert; Novotny, David; Baker-Jarvis, James; Janezic, Michael
2007-01-01
Building materials are often incorporated into complex, multilayer macrostructures that are simply not amenable to measurements using coax or waveguide sample holders. In response to this, we developed an ultra-wideband (UWB) free-field measurement system. This measurement system uses a ground-plane-based system and two TEM half-horn antennas to transmit and receive the RF signal. The material samples are placed between the antennas, and reflection and transmission measurements made. Digital signal processing techniques are then applied to minimize environmental and systematic effects. The processed data are compared to a plane-wave model to extract the material properties with optimization software based on genetic algorithms.
Hunter, Adam; Dayalan, Saravanan; De Souza, David; Power, Brad; Lorrimar, Rodney; Szabo, Tamas; Nguyen, Thu; O'Callaghan, Sean; Hack, Jeremy; Pyke, James; Nahid, Amsha; Barrero, Roberto; Roessner, Ute; Likic, Vladimir; Tull, Dedreia; Bacic, Antony; McConville, Malcolm; Bellgard, Matthew
2017-01-01
An increasing number of research laboratories and core analytical facilities around the world are developing high throughput metabolomic analytical and data processing pipelines that are capable of handling hundreds to thousands of individual samples per year, often over multiple projects, collaborations and sample types. At present, there are no Laboratory Information Management Systems (LIMS) that are specifically tailored for metabolomics laboratories that are capable of tracking samples and associated metadata from the beginning to the end of an experiment, including data processing and archiving, and which are also suitable for use in large institutional core facilities or multi-laboratory consortia as well as single laboratory environments. Here we present MASTR-MS, a downloadable and installable LIMS solution that can be deployed either within a single laboratory or used to link workflows across a multisite network. It comprises a Node Management System that can be used to link and manage projects across one or multiple collaborating laboratories; a User Management System which defines different user groups and privileges of users; a Quote Management System where client quotes are managed; a Project Management System in which metadata is stored and all aspects of project management, including experimental setup, sample tracking and instrument analysis, are defined, and a Data Management System that allows the automatic capture and storage of raw and processed data from the analytical instruments to the LIMS. MASTR-MS is a comprehensive LIMS solution specifically designed for metabolomics. It captures the entire lifecycle of a sample starting from project and experiment design to sample analysis, data capture and storage. It acts as an electronic notebook, facilitating project management within a single laboratory or a multi-node collaborative environment. This software is being developed in close consultation with members of the metabolomics research community. It is freely available under the GNU GPL v3 licence and can be accessed from, https://muccg.github.io/mastr-ms/.
NASA Astrophysics Data System (ADS)
Yussup, N.; Rahman, N. A. A.; Ibrahim, M. M.; Mokhtar, M.; Salim, N. A. A.; Soh@Shaari, S. C.; Azman, A.
2017-01-01
Neutron Activation Analysis (NAA) process has been established in Malaysian Nuclear Agency (Nuclear Malaysia) since 1980s. Most of the procedures established especially from sample registration to sample analysis are performed manually. These manual procedures carried out by the NAA laboratory personnel are time consuming and inefficient. Hence, a software to support the system automation is developed to provide an effective method to replace redundant manual data entries and produce faster sample analysis and calculation process. This paper describes the design and development of automation software for NAA process which consists of three sub-programs. The sub-programs are sample registration, hardware control and data acquisition; and sample analysis. The data flow and connection between the sub-programs will be explained. The software is developed by using National Instrument LabView development package.
The Sample Analysis at Mars Investigation and Instrument Suite
NASA Technical Reports Server (NTRS)
Mahaffy, Paul; Webster, Chris R.; Cabane, M.; Conrad, Pamela G.; Coll, Patrice; Atreya, Sushil K.; Arvey, Robert; Barciniak, Michael; Benna, Mehdi; Bleacher, L.;
2012-01-01
The Sample Analysis at Mars (SAM) investigation of the Mars Science Laboratory(MSL) addresses the chemical and isotopic composition of the atmosphere and volatilesextracted from solid samples. The SAM investigation is designed to contribute substantiallyto the mission goal of quantitatively assessing the habitability of Mars as an essentialstep in the search for past or present life on Mars. SAM is a 40 kg instrument suite locatedin the interior of MSLs Curiosity rover. The SAM instruments are a quadrupole massspectrometer, a tunable laser spectrometer, and a 6-column gas chromatograph all coupledthrough solid and gas processing systems to provide complementary information on thesame samples. The SAM suite is able to measure a suite of light isotopes and to analyzevolatiles directly from the atmosphere or thermally released from solid samples. In additionto measurements of simple inorganic compounds and noble gases SAM will conducta sensitive search for organic compounds with either thermal or chemical extraction fromsieved samples delivered by the sample processing system on the Curiosity rovers roboticarm.
Kuipers performs Water Sample Analysis
2012-05-15
ISS031-E-084619 (15 May 2012) --- After collecting samples from the Water Recovery System (WRS), European Space Agency astronaut Andre Kuipers, Expedition 31 flight engineer, processes the samples for chemical and microbial analysis in the Unity node of the International Space Station.
Kim, Hyejung; Van Hoof, Chris; Yazicioglu, Refet Firat
2011-01-01
This paper describes a mixed-signal ECG processing platform with an 12-bit ADC architecture that can adapt its sampling rate according to the input signals rate of change. This enables the sampling of ECG signals with significantly reduced data rate without loss of information. The presented adaptive sampling scheme reduces the ADC power consumption, enables the processing of ECG signals with lower power consumption, and reduces the power consumption of the radio while streaming the ECG signals. The test results show that running a CWT-based R peak detection algorithm using the adaptively sampled ECG signals consumes only 45.6 μW and it leads to 36% less overall system power consumption.
A laboratory information management system for the analysis of tritium (3H) in environmental waters.
Belachew, Dagnachew Legesse; Terzer-Wassmuth, Stefan; Wassenaar, Leonard I; Klaus, Philipp M; Copia, Lorenzo; Araguás, Luis J Araguás; Aggarwal, Pradeep
2018-07-01
Accurate and precise measurements of low levels of tritium ( 3 H) in environmental waters are difficult to attain due to complex steps of sample preparation, electrolytic enrichment, liquid scintillation decay counting, and extensive data processing. We present a Microsoft Access™ relational database application, TRIMS (Tritium Information Management System) to assist with sample and data processing of tritium analysis by managing the processes from sample registration and analysis to reporting and archiving. A complete uncertainty propagation algorithm ensures tritium results are reported with robust uncertainty metrics. TRIMS will help to increase laboratory productivity and improve the accuracy and precision of 3 H assays. The software supports several enrichment protocols and LSC counter types. TRIMS is available for download at no cost from the IAEA at www.iaea.org/water. Copyright © 2018 Elsevier Ltd. All rights reserved.
Towards a Mobile Ecogenomic sensor: the Third Generation Environmental Sample Processor (3G-ESP).
NASA Astrophysics Data System (ADS)
Birch, J. M.; Pargett, D.; Jensen, S.; Roman, B.; Preston, C. M.; Ussler, W.; Yamahara, K.; Marin, R., III; Hobson, B.; Zhang, Y.; Ryan, J. P.; Scholin, C. A.
2016-02-01
Researchers are increasingly using one or more autonomous platforms to characterize ocean processes that change in both space and time. Conceptually, studying processes that change quickly both spatially and temporally seems relatively straightforward. One needs to sample in many locations synoptically over time, or follow a coherent water mass and sample it repeatedly. However, implementing either approach presents many challenges. For example, acquiring samples over days to weeks far from shore, without human intervention, requires multiple systems to work together seamlessly, and the level of autonomy, navigation and communications needed to conduct the work exposes the complexity of these requirements. We are addressing these challenges by developing a new generation of robotic systems that are primarily aimed at studies of microbial-mediated processes. As a step towards realizing this new capability, we have taken lessons learned from our second-generation Environmental Sample Processor (2G-ESP), a robotic microbiology "lab-in-a-can" and have re-engineered the system for use on a Tethys-class Long Range AUV (LRAUV). The new instrument is called the third-generation ESP (3G-ESP), and its integration with the LRAUV provides mobility and a persistent presence not seen before in microbial oceanography. The 3G-ESP autonomously filters a water sample and then either preserves that material for eventual return to a laboratory, or processes the sample in real-time for further downstream molecular analytical analyses. The 3G ESP modularizes hardware needed for the collection and preparation of a sample from subsequent molecular analyses by the use of self-contained "cartridges". Cartridges currently come in two forms: one for the preservation of a sample, and the other for onboard homogenization and handoff for downstream processing via one or more analytical devices. The 3G-ESP is designed as a stand-alone instrument, and thus could be deployed on a variety of platforms. This presentation will focus on results from early deployments of the prototype 3G-ESP/LRAUV, the challenges encountered in cartridge design, ESP/LRAUV integration, and operational capabilities that show the potential of mobile, ecogenomic sensors in the ocean sciences.
NASA Technical Reports Server (NTRS)
1994-01-01
The ChemScan UV-6100 is a spectrometry system originally developed by Biotronics Technologies, Inc. under a Small Business Innovation Research (SBIR) contract. It is marketed to the water and wastewater treatment industries, replacing "grab sampling" with on-line data collection. It analyzes the light absorbance characteristics of a water sample, simultaneously detects hundreds of individual wavelengths absorbed by chemical substances in a process solution, and quantifies the information. Spectral data is then processed by ChemScan analyzer and compared with calibration files in the system's memory in order to calculate concentrations of chemical substances that cause UV light absorbance in specific patterns. Monitored substances can be analyzed for quality and quantity. Applications include detection of a variety of substances, and the information provided enables an operator to control a process more efficiently.
A Real-Time Image Acquisition And Processing System For A RISC-Based Microcomputer
NASA Astrophysics Data System (ADS)
Luckman, Adrian J.; Allinson, Nigel M.
1989-03-01
A low cost image acquisition and processing system has been developed for the Acorn Archimedes microcomputer. Using a Reduced Instruction Set Computer (RISC) architecture, the ARM (Acorn Risc Machine) processor provides instruction speeds suitable for image processing applications. The associated improvement in data transfer rate has allowed real-time video image acquisition without the need for frame-store memory external to the microcomputer. The system is comprised of real-time video digitising hardware which interfaces directly to the Archimedes memory, and software to provide an integrated image acquisition and processing environment. The hardware can digitise a video signal at up to 640 samples per video line with programmable parameters such as sampling rate and gain. Software support includes a work environment for image capture and processing with pixel, neighbourhood and global operators. A friendly user interface is provided with the help of the Archimedes Operating System WIMP (Windows, Icons, Mouse and Pointer) Manager. Windows provide a convenient way of handling images on the screen and program control is directed mostly by pop-up menus.
Vogeser, Michael; Spöhrer, Ute
2006-01-01
Liquid chromatography tandem-mass spectrometry (LC-MS/MS) is an efficient technology for routine determination of immunosuppressants in whole blood; however, time-consuming manual sample preparation remains a significant limitation of this technique. Using a commercially available robotic pipetting system (Tecan Freedom EVO), we developed an automated sample-preparation protocol for quantification of tacrolimus in whole blood by LC-MS/MS. Barcode reading, sample resuspension, transfer of whole blood aliquots into a deep-well plate, addition of internal standard solution, mixing, and protein precipitation by addition of an organic solvent is performed by the robotic system. After centrifugation of the plate, the deproteinized supernatants are submitted to on-line solid phase extraction, using column switching prior to LC-MS/MS analysis. The only manual actions within the entire process are decapping of the tubes, and transfer of the deep-well plate from the robotic system to a centrifuge and finally to the HPLC autosampler. Whole blood pools were used to assess the reproducibility of the entire analytical system for measuring tacrolimus concentrations. A total coefficient of variation of 1.7% was found for the entire automated analytical process (n=40; mean tacrolimus concentration, 5.3 microg/L). Close agreement between tacrolimus results obtained after manual and automated sample preparation was observed. The analytical system described here, comprising automated protein precipitation, on-line solid phase extraction and LC-MS/MS analysis, is convenient and precise, and minimizes hands-on time and the risk of mistakes in the quantification of whole blood immunosuppressant concentrations compared to conventional methods.
Packaging Considerations for Biopreservation
Woods, Erik J.; Thirumala, Sreedhar
2011-01-01
Summary The packaging system chosen for biopreservation is critical for many reasons. An ideal biopreservation container system must provide for closure integrity, sample stability and ready access to the preserved material. This means the system needs to be hermetically sealed to ensure integrity of the specimen is maintained throughout processing, storage and distribution; the system must remain stable over long periods of time as many biobanked samples may be stored indefinitely; and functionally closed access systems must be used to avoid contamination upon sample withdraw. This study reviews the suitability of a new commercially available vial configuration container utilizing blood bag style closure and access systems that can be hermetically sealed and remain stable through cryopreservation and biobanking procedures. This vial based systems allow for current good manufacturing/tissue practice (cGTP) requirements during processing of samples and may provide the benefit of ease of delivery by a care giver. In this study, the CellSeal® closed system cryovial was evaluated and compared to standard screw cap vials. The CellSeal system was evaluated for durability, closure integrity through transportation and maintenance of functional viability of a cryopreserved mesenchymal stem cell model. The results of this initial proof-of-concept study indicated that the CellSeal vials are highly suitable for biopreservation and biobanking, and provide a suitable container system for clinical and commercial cell therapy products frozen in small volumes. PMID:21566715
A flexible system to capture sample vials in a storage box - the box vial scanner.
Nowakowski, Steven E; Kressin, Kenneth R; Deick, Steven D
2009-01-01
Tracking sample vials in a research environment is a critical task and doing so efficiently can have a large impact on productivity, especially in high volume laboratories. There are several challenges to automating the capture process, including the variety of containers used to store samples. We developed a fast and robust system to capture the location of sample vials being placed in storage that allows the laboratories the flexibility to use sample containers of varying dimensions. With a single scan, this device captures the box identifier, the vial identifier and the location of each vial within a freezer storage box. The sample vials are tracked through a barcode label affixed to the cap while the boxes are tracked by a barcode label on the side of the box. Scanning units are placed at the point of use and forward data to a sever application for processing the scanned data. Scanning units consist of an industrial barcode reader mounted in a fixture positioning the box for scanning and providing lighting during the scan. The server application transforms the scan data into a list of storage locations holding vial identifiers. The list is then transferred to the laboratory database. The box vial scanner captures the IDs and location information for an entire box of sample vials into the laboratory database in a single scan. The system accommodates a wide variety of vials sizes by inserting risers under the sample box and a variety of storage box layouts are supported via the processing algorithm on the server.
Kukreti, B M; Sharma, G K
2012-05-01
Accurate and speedy estimations of ppm range uranium and thorium in the geological and rock samples are most useful towards ongoing uranium investigations and identification of favorable radioactive zones in the exploration field areas. In this study with the existing 5 in. × 4 in. NaI(Tl) detector setup, prevailing background and time constraints, an enhanced geometrical setup has been worked out to improve the minimum detection limits for primordial radioelements K(40), U(238) and Th(232). This geometrical setup has been integrated with the newly introduced, digital signal processing based MCA system for the routine spectrometric analysis of low concentration rock samples. Stability performance, during the long counting hours, for digital signal processing MCA system and its predecessor NIM bin based MCA system has been monitored, using the concept of statistical process control. Monitored results, over a time span of few months, have been quantified in terms of spectrometer's parameters such as Compton striping constants and Channel sensitivities, used for evaluating primordial radio element concentrations (K(40), U(238) and Th(232)) in geological samples. Results indicate stable dMCA performance, with a tendency of higher relative variance, about mean, particularly for Compton stripping constants. Copyright © 2012 Elsevier Ltd. All rights reserved.
ROMPS critical design review. Volume 1: Hardware
NASA Technical Reports Server (NTRS)
Dobbs, M. E.
1992-01-01
Topics concerning the Robot-Operated Material Processing in Space (ROMPS) Program are presented in viewgraph form and include the following: a systems overview; servocontrol and servomechanisms; testbed and simulation results; system V controller; robot module; furnace module; SCL experiment supervisor; SCL script sample processing control; SCL experiment supervisor fault handling; block diagrams; hitchhiker interfaces; battery systems; watchdog timers; mechanical/thermal systems; and fault conditions and recovery.
Flexible automated approach for quantitative liquid handling of complex biological samples.
Palandra, Joe; Weller, David; Hudson, Gary; Li, Jeff; Osgood, Sarah; Hudson, Emily; Zhong, Min; Buchholz, Lisa; Cohen, Lucinda H
2007-11-01
A fully automated protein precipitation technique for biological sample preparation has been developed for the quantitation of drugs in various biological matrixes. All liquid handling during sample preparation was automated using a Hamilton MicroLab Star Robotic workstation, which included the preparation of standards and controls from a Watson laboratory information management system generated work list, shaking of 96-well plates, and vacuum application. Processing time is less than 30 s per sample or approximately 45 min per 96-well plate, which is then immediately ready for injection onto an LC-MS/MS system. An overview of the process workflow is discussed, including the software development. Validation data are also provided, including specific liquid class data as well as comparative data of automated vs manual preparation using both quality controls and actual sample data. The efficiencies gained from this automated approach are described.
Theory on data processing and instrumentation. [remote sensing
NASA Technical Reports Server (NTRS)
Billingsley, F. C.
1978-01-01
A selection of NASA Earth observations programs are reviewed, emphasizing hardware capabilities. Sampling theory, noise and detection considerations, and image evaluation are discussed for remote sensor imagery. Vision and perception are considered, leading to numerical image processing. The use of multispectral scanners and of multispectral data processing systems, including digital image processing, is depicted. Multispectral sensing and analysis in application with land use and geographical data systems are also covered.
State Tracking and Fault Diagnosis for Dynamic Systems Using Labeled Uncertainty Graph.
Zhou, Gan; Feng, Wenquan; Zhao, Qi; Zhao, Hongbo
2015-11-05
Cyber-physical systems such as autonomous spacecraft, power plants and automotive systems become more vulnerable to unanticipated failures as their complexity increases. Accurate tracking of system dynamics and fault diagnosis are essential. This paper presents an efficient state estimation method for dynamic systems modeled as concurrent probabilistic automata. First, the Labeled Uncertainty Graph (LUG) method in the planning domain is introduced to describe the state tracking and fault diagnosis processes. Because the system model is probabilistic, the Monte Carlo technique is employed to sample the probability distribution of belief states. In addition, to address the sample impoverishment problem, an innovative look-ahead technique is proposed to recursively generate most likely belief states without exhaustively checking all possible successor modes. The overall algorithms incorporate two major steps: a roll-forward process that estimates system state and identifies faults, and a roll-backward process that analyzes possible system trajectories once the faults have been detected. We demonstrate the effectiveness of this approach by applying it to a real world domain: the power supply control unit of a spacecraft.
High frequency signal acquisition and control system based on DSP+FPGA
NASA Astrophysics Data System (ADS)
Liu, Xiao-qi; Zhang, Da-zhi; Yin, Ya-dong
2017-10-01
This paper introduces a design and implementation of high frequency signal acquisition and control system based on DSP + FPGA. The system supports internal/external clock and internal/external trigger sampling. It has a maximum sampling rate of 400MBPS and has a 1.4GHz input bandwidth for the ADC. Data can be collected continuously or periodically in systems and they are stored in DDR2. At the same time, the system also supports real-time acquisition, the collected data after digital frequency conversion and Cascaded Integrator-Comb (CIC) filtering, which then be sent to the CPCI bus through the high-speed DSP, can be assigned to the fiber board for subsequent processing. The system integrates signal acquisition and pre-processing functions, which uses high-speed A/D, high-speed DSP and FPGA mixed technology and has a wide range of uses in data acquisition and recording. In the signal processing, the system can be seamlessly connected to the dedicated processor board. The system has the advantages of multi-selectivity, good scalability and so on, which satisfies the different requirements of different signals in different projects.
Automatic bio-sample bacteria detection system
NASA Technical Reports Server (NTRS)
Chappelle, E. W.; Colburn, M.; Kelbaugh, B. N.; Picciolo, G. L.
1971-01-01
Electromechanical device analyzes urine specimens in 15 minutes and processes one sample per minute. Instrument utilizes bioluminescent reaction between luciferase-luciferin mixture and adenosine triphosphate (ATP) to determine number of bacteria present in the sample. Device has potential application to analysis of other body fluids.
Wessendorf, Kurt O.; Kemper, Dale A.
2003-06-03
A very low power analog pulse processing system implemented as an ASIC useful for processing signals from radiation detectors, among other things. The system incorporates the functions of a charge sensitive amplifier, a shaping amplifier, a peak sample and hold circuit, and, optionally, an analog to digital converter and associated drivers.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-10-25
... FURTHER INFORMATION CONTACT: For Legal Aspects: Alan C. Cohen, Penalties Branch, Regulations and Rulings... claimant to resolve defects. It is recognized that in some cases the sampling will be so flawed it cannot... companies' internal processes and systems during the application process. ISA members are companies with...
NASA Astrophysics Data System (ADS)
Kozlova, Tatiana; Karol Seweryn, D..; Grygorczuk, Jerzy; Kozlov, Oleg
The sample return missions have made a very significant progress to understanding of geology, the extra-terrestrial materials, processes occurring on surface and subsurface level, as well as of interactions between such materials and mechanisms operating there. The various sample return missions in the past (e.g. Apollo missions, Luna missions, Hayabusa mission) have provided scientists with samples of extra-terrestrial materials allowing to discover answers to critical scientific questions concerning the origin and evolution of the Solar System. Several new missions are currently planned: sample return missions, e.g Russian Luna-28, ESA Phootprint and MarcoPolo-R as well as both robotic and manned exploration missions to the Moon and Mars. One of the key challenges in such missions is the reliable sampling process which can be achieved by using many different techniques, e.g. static excavating technique (scoop), core drilling, sampling using dynamic mechanisms (penetrators), brushes and pneumatic systems. The effectiveness of any sampling strategy depends on many factors, including the required sample size, the mechanical and chemical soil properties (cohesive, hard or porous regolith, stones), the environment conditions (gravity, temperature, pressure, radiation). Many sampling mechanism have been studied, designed and built in the past, two techniques to collect regolith samples were chosen for the Phobos-Grunt mission. The proposed system consisted of a robotic arm with a 1,2m reach beyond the lander (IKI RAN); a tubular sampling device designed for collecting both regolith and small rock fragments (IKI RAN); the CHOMIK device (CBK PAN) - the low velocity penetrator with a single-sample container for collecting samples from the rocky surface. The functional tests were essential step in robotic arm, sampling device and CHOMIK device development process in the frame of Phobos-Grunt mission. Three major results were achieved: (i) operation scenario for autonomous sampling; (ii) technical characteristics of both devices, i.e. progress cycles of CHOMIK device in different materials and torque in the manipulator joints during sampling operations; (iii) confirmation of applicability of both devices to perform such type of tasks. The phases in operational scenario were prepared to meet mission and system requirements mainly connected with: (i) environment (near zero gravity, vacuum, dust), (ii) safety and (iii) to avoid common operation of both devices at the same time.
77 FR 26507 - Application(s) for Duty-Free Entry of Scientific Instruments
Federal Register 2010, 2011, 2012, 2013, 2014
2012-05-04
...: FEI Company, Czech Republic. Intended Use: The instrument will be used for research on primitive solar system materials extracted from meteorites as well as on samples from NASA sample return missions, such... chemical origin of the solar system and the processes by which its small bodies evolved. Justification for...
Lam, C K; Zhang, Y; Busch, M A; Busch, K W
1993-06-01
A new sample introduction system for the analysis of continuously flowing liquid streams by flame infrared-emission (FIRE) spectrometry has been developed. The system uses a specially designed purge cell to strip dissolved CO(2) from solution into a hydrogen gas stream that serves as the fuel for a hydrogen/air flame. Vibrationally excited CO(2) molecules present in the flame are monitored with a simple infrared filter (4.4 mum) photometer. The new system can be used to introduce analytes as a continuous liquid stream (process analysis mode) or on a discrete basis by sample injection (flow injection analysis mode). The key to the success of the method is the new purge-cell design. The small internal volume of the cell minimizes problems associated with purge-cell clean-out and produces sharp, reproducible signals. Spent analytical solution is continuously drained from the cell, making cell disconnection and cleaning between samples unnecessary. Under the conditions employed in this study, samples could be analyzed at a maximum rate of approximately 60/h. The new sample introduction system was successfully tested in both a process analysis- and a flow injection analysis mode for the determination of total inorganic carbon in Waco tap water. For the first time, flame infrared-emission spectrometry was successfully extended to non-volatile organic compounds by using chemical pretreatment with peroxydisulfate in the presence of silver ion to convert the analytes into dissolved carbon dioxide, prior to purging and detection by the FIRE radiometer. A test of the peroxydisulfate/Ag(+) reaction using six organic acids and five sugars indicated that all 11 compounds were oxidized to nearly the same extent. Finally, the new sample introduction system was used in conjunction with a simple filter FIRE radiometer as a detection system in ion-exchange high-performance liquid chromatography. Ion-exchange chromatograms are shown for two aqueous mixtures, one containing six organic acids and the second containing six mono-, di-, and trisaccharides.
Shanmugam, Akshaya; Usmani, Mohammad; Mayberry, Addison; Perkins, David L; Holcomb, Daniel E
2018-01-01
Miniaturized imaging devices have pushed the boundaries of point-of-care imaging, but existing mobile-phone-based imaging systems do not exploit the full potential of smart phones. This work demonstrates the use of simple imaging configurations to deliver superior image quality and the ability to handle a wide range of biological samples. Results presented in this work are from analysis of fluorescent beads under fluorescence imaging, as well as helminth eggs and freshwater mussel larvae under white light imaging. To demonstrate versatility of the systems, real time analysis and post-processing results of the sample count and sample size are presented in both still images and videos of flowing samples.
Digital Curation of Earth Science Samples Starts in the Field
NASA Astrophysics Data System (ADS)
Lehnert, K. A.; Hsu, L.; Song, L.; Carter, M. R.
2014-12-01
Collection of physical samples in the field is an essential part of research in the Earth Sciences. Samples provide a basis for progress across many disciplines, from the study of global climate change now and over the Earth's history, to present and past biogeochemical cycles, to magmatic processes and mantle dynamics. The types of samples, methods of collection, and scope and scale of sampling campaigns are highly diverse, ranging from large-scale programs to drill rock and sediment cores on land, in lakes, and in the ocean, to environmental observation networks with continuous sampling, to single investigator or small team expeditions to remote areas around the globe or trips to local outcrops. Cyberinfrastructure for sample-related fieldwork needs to cater to the different needs of these diverse sampling activities, aligning with specific workflows, regional constraints such as connectivity or climate, and processing of samples. In general, digital tools should assist with capture and management of metadata about the sampling process (location, time, method) and the sample itself (type, dimension, context, images, etc.), management of the physical objects (e.g., sample labels with QR codes), and the seamless transfer of sample metadata to data systems and software relevant to the post-sampling data acquisition, data processing, and sample curation. In order to optimize CI capabilities for samples, tools and workflows need to adopt community-based standards and best practices for sample metadata, classification, identification and registration. This presentation will provide an overview and updates of several ongoing efforts that are relevant to the development of standards for digital sample management: the ODM2 project that has generated an information model for spatially-discrete, feature-based earth observations resulting from in-situ sensors and environmental samples, aligned with OGC's Observation & Measurements model (Horsburgh et al, AGU FM 2014); implementation of the IGSN (International Geo Sample Number) as a globally unique sample identifier via a distributed system of allocating agents and a central registry; and the EarthCube Research Coordination Network iSamplES (Internet of Samples in the Earth Sciences) that aims to improve sharing and curation of samples through the use of CI.
Improvement in the stability of serum samples stored in an automated refrigerated module.
Parra-Robert, Marina; Rico-Santana, Naira; Alcaraz-Quiles, José; Sandalinas, Silvia; Fernández, Esther; Falcón, Isabel; Pérez-Riedweg, Margarita; Bedini, Josep Lluís
2016-12-01
In clinical laboratories it is necessary to know for how long the analytes are stable in the samples with specific storage conditions. Our laboratory has implemented the new Aptio Automation System (AAS) (Siemens Healthcare Diagnostics) where the analyzed samples are stored in a refrigerated storage module (RSM) after being sealed. The aim of the study was to evaluate the stability of serum samples with the AAS and comparing the results with a previous study using a conventional refrigerated system. Serum samples from a total of 50 patients were collected and for each of them 27 biochemical analytes were analyzed. The samples were divided in 5 sets of 10 samples. Each set was re-analyzed at one of the following times: 24, 48, 72, 96 and 120h. Stability was evaluated according to the Total Limit of Change (TLC) criteria, which combine both analytical and biologic variation. A total of 26 out of 27 analytes were stable at the end of the study according to TLC criteria. Lactate dehydrogenase was not stable at 48h observing a decrease in its concentration until the end of the study. In the previous study (conventional storage system) 9 biochemical analytes were not stable with an increase of their levels due to the evaporation process. The RSM connected to the AAS improves the stability of serum samples. This system avoids the evaporation process due to the sealing of samples and allows better control of the samples during their storage. Copyright © 2016 The Canadian Society of Clinical Chemists. Published by Elsevier Inc. All rights reserved.
Indoor and Outdoor Mobile Mapping Systems for Architectural Surveys
NASA Astrophysics Data System (ADS)
Campi, M.; di Luggo, A.; Monaco, S.; Siconolfi, M.; Palomba, D.
2018-05-01
This paper presents the results of architectural surveys carried out with mobile mapping systems. The data acquired through different instruments for both indoor and outdoor surveying are analyzed and compared. The study sample shows what is required for an acquisition in a dynamic mode indicating the criteria for the creation of a georeferenced network for indoor spaces, as well as the operational processes concerning data capture, processing, and management. The differences between a dynamic and static scan have been evaluated, with a comparison being made with the aerial photogrammetric survey of the same sample.
Containerless processing of undercooled melts
NASA Technical Reports Server (NTRS)
Perepezko, J. H.
1993-01-01
The investigation focused on the control of microstructural evolution in Mn-Al, Fe-Ni, Ni-V, and Au-Pb-Sb alloys through the high undercooling levels provided by containerless processing, and provided fundamental new information on the control of nucleation. Solidification analysis was conducted by means of thermal analysis, x-ray diffraction, and metallographic characterization on samples processed in a laboratory scale drop tube system. The Mn-Al alloy system offers a useful model system with the capability of phase separation on an individual particle basis, thus permitting a more complete understanding of the operative kinetics and the key containerless processing variables. This system provided the opportunity of analyzing the nucleation rate as a function of processing conditions and allowed for the quantitative assessment of the relevant processing parameters. These factors are essential in the development of a containerless processing model which has a predictive capability. Similarly, Ni-V is a model system that was used to study duplex partitionless solidification, which is a structure possible only in high under cooling solidification processes. Nucleation kinetics for the competing bcc and fcc phases were studied to determine how this structure can develop and the conditions under which it may occur. The Fe-Ni alloy system was studied to identify microstructural transitions with controlled variations in sample size and composition during containerless solidification. This work was forwarded to develop a microstructure map which delineates regimes of structural evolution and provides a unified analysis of experimental observations. The Au-Pb-Sb system was investigated to characterize the thermodynamic properties of the undercooled liquid phase and to characterize the glass transition under a variety of processing conditions. By analyzing key containerless processing parameters in a ground based drop tube study, a carefully designed flight experiment may be planned to utilize the extended duration microgravity conditions of orbiting spacecraft.
Barium isotope abundances in meteorites and their implications for early Solar System evolution
NASA Astrophysics Data System (ADS)
Bermingham, K. R.; Mezger, K.; Scherer, E. E.; Horan, M. F.; Carlson, R. W.; Upadhyay, D.; Magna, T.; Pack, A.
2016-02-01
Several nucleosynthetic processes contributed material to the Solar System, but the relative contributions of each process, the timing of their input into the solar nebula, and how well these components were homogenized in the solar nebula remain only partially constrained. The Ba isotope system is particularly useful in addressing these issues because Ba isotopes are synthesized via three nucleosynthetic processes (s-, r-, p-process). In this study, high precision Ba isotope analyses of 22 different whole rock chondrites and achondrites (carbonaceous chondrites, ordinary chondrites, enstatite chondrites, Martian meteorites, and eucrites) were performed to constrain the distribution of Ba isotopes on the regional scale in the Solar System. A melting method using aerodynamic levitation and CO2-laser heating was used to oxidize SiC, a primary carrier of Ba among presolar grains in carbonaceous chondrites. Destruction of these grains during the fusion process enabled the complete digestion of these samples. The Ba isotope data presented here are thus the first for which complete dissolution of the bulk meteorite samples was certain. Enstatite chondrites, ordinary chondrites, and all achondrites measured here possess Ba isotope compositions that are not resolved from the terrestrial composition. Barium isotope anomalies are evident in most of the carbonaceous chondrites analyzed, but the 135Ba anomalies are generally smaller than previously reported for similarly sized splits of CM2 meteorites. Variation in the size of the 135Ba anomaly is also apparent in fused samples from the same parent body (e.g., CM2 meteorites) and in different pieces from the same meteorite (e.g., Orgueil, CI). Here, we investigate the potential causes of variability in 135Ba, including the contribution of radiogenic 135Ba from the decay of 135Cs and incomplete homogenization of the presolar components on the <0.8 g sample scale.
A high speed implementation of the random decrement algorithm
NASA Technical Reports Server (NTRS)
Kiraly, L. J.
1982-01-01
The algorithm is useful for measuring net system damping levels in stochastic processes and for the development of equivalent linearized system response models. The algorithm works by summing together all subrecords which occur after predefined threshold level is crossed. The random decrement signature is normally developed by scanning stored data and adding subrecords together. The high speed implementation of the random decrement algorithm exploits the digital character of sampled data and uses fixed record lengths of 2(n) samples to greatly speed up the process. The contributions to the random decrement signature of each data point was calculated only once and in the same sequence as the data were taken. A hardware implementation of the algorithm using random logic is diagrammed and the process is shown to be limited only by the record size and the threshold crossing frequency of the sampled data. With a hardware cycle time of 200 ns and 1024 point signature, a threshold crossing frequency of 5000 Hertz can be processed and a stably averaged signature presented in real time.
LISP based simulation generators for modeling complex space processes
NASA Technical Reports Server (NTRS)
Tseng, Fan T.; Schroer, Bernard J.; Dwan, Wen-Shing
1987-01-01
The development of a simulation assistant for modeling discrete event processes is presented. Included are an overview of the system, a description of the simulation generators, and a sample process generated using the simulation assistant.
An advanced material science payload for GAS
NASA Technical Reports Server (NTRS)
Joensson, R.; Wallin, S.; Loeth, K.
1986-01-01
The aim of the experiment is to study solidification of different compositions of lead-tin. The weight of the material is quite high: 8 kilograms. Nearly 10% of the payload is sample weight. The dendritic growth and the effect of the absence of natural convection are of particular interest. The results from the flight processed samples will be compared with results from Earth processed samples in order to investigate the influence of the natural convection on the solidification process. The power systems, heat storage and rejection, and mechanical support are discussed in relationship to the scientific requirements.
Process control using fiber optics and Fourier transform infrared spectroscopy
NASA Astrophysics Data System (ADS)
Kemsley, E. K.; Wilson, Reginald H.
1992-03-01
A process control system has been constructed using optical fibers interfaced to a Fourier transform infrared (FT-IR) spectrometer, to achieve remote spectroscopic analysis of food samples during processing. The multichannel interface accommodates six fibers, allowing the sequential observation of up to six samples. Novel fiber-optic sampling cells have been constructed, including transmission and attenuated total reflectance (ATR) designs. Different fiber types have been evaluated; in particular, plastic clad silica (PCS) and zirconium fluoride fibers. Processes investigated have included the dilution of fruit juice concentrate, and the addition of alcohol to fruit syrup. Suitable algorithms have been written which use the results of spectroscopic measurements to control and monitor the course of each process, by actuating devices such as valves and switches.
Preparation and recrystallization behavior of spray-dried co-amorphous naproxen-indomethacin.
Beyer, Andreas; Radi, Lydia; Grohganz, Holger; Löbmann, Korbinian; Rades, Thomas; Leopold, Claudia S
2016-07-01
To improve the dissolution properties and the physical stability of amorphous active pharmaceutical ingredients, small molecule stabilizing agents may be added to prepare co-amorphous systems. The objective of the study was to investigate if spray-drying allows the preparation of co-amorphous drug-drug systems such as naproxen-indomethacin and to examine the influence of the process conditions on the resulting initial sample crystallinity and the recrystallization behavior of the drug(s). For this purpose, the process parameters inlet temperature and pump feed rate were varied according to a 2(2) factorial design and the obtained samples were analyzed with X-ray powder diffractometry and Fourier-transformed infrared spectroscopy. Evaluation of the data revealed that the preparation of fully amorphous samples could be achieved depending on the process conditions. The resulting recrystallization behavior of the samples, such as the total recrystallization rate, the individual recrystallization rates of naproxen and indomethacin as well as the polymorphic form of indomethacin that was formed were influenced by these process conditions. For initially amorphous samples, it was found that naproxen and indomethacin recrystallized almost simultaneously, which supports the theory of formation of drug-drug heterodimers in the co-amorphous phase. Copyright © 2016 Elsevier B.V. All rights reserved.
Wang, Tong; Gao, Huijun; Qiu, Jianbin
2016-02-01
This paper investigates the multirate networked industrial process control problem in double-layer architecture. First, the output tracking problem for sampled-data nonlinear plant at device layer with sampling period T(d) is investigated using adaptive neural network (NN) control, and it is shown that the outputs of subsystems at device layer can track the decomposed setpoints. Then, the outputs and inputs of the device layer subsystems are sampled with sampling period T(u) at operation layer to form the index prediction, which is used to predict the overall performance index at lower frequency. Radial basis function NN is utilized as the prediction function due to its approximation ability. Then, considering the dynamics of the overall closed-loop system, nonlinear model predictive control method is proposed to guarantee the system stability and compensate the network-induced delays and packet dropouts. Finally, a continuous stirred tank reactor system is given in the simulation part to demonstrate the effectiveness of the proposed method.
Yan, Zhinong; Vorst, Keith L; Zhang, Lei; Ryser, Elliot T
2007-05-01
A novel one-ply composite tissue (CT) method using the Soleris (formerly BioSys) optical analysis system was compared with the conventional U.S. Department of Agriculture (USDA) environmental sponge enrichment method for recovery of Listeria from food contact surfaces and poultry-processing environments. Stainless steel and high-density polyethylene plates were inoculated to contain a six-strain L. monocytogenes cocktail at 10(4), 10(2), and 10 CFU per plate, whereas samples from naturally contaminated surfaces and floor drains from a poultry-processing facility were collected with CTs and environmental sponges. CT samples were transferred into Soleris system vials, and presumptive-positive samples were further confirmed. Sponge samples were processed for Listeria using the USDA culture method. L. monocytogenes recovery rates from inoculated stainless steel and polyethylene surfaces were then compared for the two methods in terms of sensitivity, specificity, and positive and negative predictive values. No significant differences (P > 0.05) were found between the two methods for recovery of L. monocytogenes from any of the inoculated stainless steel and polyethylene surfaces or environmental samples. Sensitivity, specificity, and overall accuracy of the CT-Soleris for recovery of Listeria from environmental samples were 83, 97, and 95%, respectively. Listeria was detected 2 to 3 days sooner with the CT-Soleris method than with the USDA culture method, thus supporting the increased efficacy of this new protocol for environmental sampling.
NASA Astrophysics Data System (ADS)
Bolhuis, Peter
Important reaction-diffusion processes, such as biochemical networks in living cells, or self-assembling soft matter, span many orders in length and time scales. In these systems, the reactants' spatial dynamics at mesoscopic length and time scales of microns and seconds is coupled to the reactions between the molecules at microscopic length and time scales of nanometers and milliseconds. This wide range of length and time scales makes these systems notoriously difficult to simulate. While mean-field rate equations cannot describe such processes, the mesoscopic Green's Function Reaction Dynamics (GFRD) method enables efficient simulation at the particle level provided the microscopic dynamics can be integrated out. Yet, many processes exhibit non-trivial microscopic dynamics that can qualitatively change the macroscopic behavior, calling for an atomistic, microscopic description. The recently developed multiscale Molecular Dynamics Green's Function Reaction Dynamics (MD-GFRD) approach combines GFRD for simulating the system at the mesocopic scale where particles are far apart, with microscopic Molecular (or Brownian) Dynamics, for simulating the system at the microscopic scale where reactants are in close proximity. The association and dissociation of particles are treated with rare event path sampling techniques. I will illustrate the efficiency of this method for patchy particle systems. Replacing the microscopic regime with a Markov State Model avoids the microscopic regime completely. The MSM is then pre-computed using advanced path-sampling techniques such as multistate transition interface sampling. I illustrate this approach on patchy particle systems that show multiple modes of binding. MD-GFRD is generic, and can be used to efficiently simulate reaction-diffusion systems at the particle level, including the orientational dynamics, opening up the possibility for large-scale simulations of e.g. protein signaling networks.
Zhu, Li-Wei; Yang, Xue-Mei; Xu, Xiao-Qin; Xu, Jian; Lu, Huang-Jun; Yan, Li-Xing
2008-10-01
This study was aimed to analyze the results of false positive reaction in bacterial detection of blood samples with BacT/ALERT 3D system, to evaluate the specificity of this system, and to decrease the false positive reaction. Each reaction flasks in past five years were processed for bacteria isolation and identification. When the initial cultures were positive, the remaining samples and the corresponding units were recultured if still available. 11395 blood samples were detected. It is worthy of note that the incubator temperature should be stabilized, avoiding fluctuation; when the cultures were alarmed, the reaction flasks showed be kept some hours for further incubation so as to trace a sharply increasing signal to support the judgement of true bacterial growth. The results indicated that 122 samples (1.07%) wee positive at initial culture, out of them 107 samples (88.7%) were found bacterial, and 15 samples (12.3%) were found nothing. The detection curves of positive samples resulted from bacterial growth showed ascent. In conclusion, maintenance of temperature stability and avoidance of temperature fluctuation in incubator could decrease the occurrence of false-positive reaction in detection process. The reaction flasks with positive results at initial culture should be recultured, and whether existence of a sharply ascending logarilhimic growth phase in bacterial growth curve should be further detected, which are helpful to distinguish false-positive reactions from true positive, and thus increase the specificity of the BacT/ALERT system.
40 CFR 65.164 - Performance test and flare compliance determination notifications and reports.
Code of Federal Regulations, 2010 CFR
2010-07-01
... PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) CONSOLIDATED FEDERAL AIR RULE Closed Vent Systems, Control Devices, and Routing to a Fuel Gas System or a Process § 65.164 Performance test and flare... complete test report shall include a brief process description, sampling site description, description of...
Stochastic stability properties of jump linear systems
NASA Technical Reports Server (NTRS)
Feng, Xiangbo; Loparo, Kenneth A.; Ji, Yuandong; Chizeck, Howard J.
1992-01-01
Jump linear systems are defined as a family of linear systems with randomly jumping parameters (usually governed by a Markov jump process) and are used to model systems subject to failures or changes in structure. The authors study stochastic stability properties in jump linear systems and the relationship among various moment and sample path stability properties. It is shown that all second moment stability properties are equivalent and are sufficient for almost sure sample path stability, and a testable necessary and sufficient condition for second moment stability is derived. The Lyapunov exponent method for the study of almost sure sample stability is discussed, and a theorem which characterizes the Lyapunov exponents of jump linear systems is presented.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kirkham, R.; Siddons, D.; Dunn, P.A.
2010-06-23
The Maia detector system is engineered for energy dispersive x-ray fluorescence spectroscopy and elemental imaging at photon rates exceeding 10{sup 7}/s, integrated scanning of samples for pixel transit times as small as 50 {micro}s and high definition images of 10{sup 8} pixels and real-time processing of detected events for spectral deconvolution and online display of pure elemental images. The system developed by CSIRO and BNL combines a planar silicon 384 detector array, application-specific integrated circuits for pulse shaping and peak detection and sampling and optical data transmission to an FPGA-based pipelined, parallel processor. This paper describes the system and themore » underpinning engineering solutions.« less
Cuffney, T.F.; Gurtz, M.E.; Meador, M.R.
1993-01-01
Benthic invertebrate samples are collected as part of the U.S. Geological Survey's National Water-Quality Assessment Program. This is a perennial, multidisciplinary program that integrates biological, physical, and chemical indicators of water quality to evaluate status and trends and to develop an understanding of the factors controlling observed water quality. The Program examines water quality in 60 study units (coupled ground- and surface-water systems) that encompass most of the conterminous United States and parts of Alaska and Hawaii. Study-unit teams collect and process qualitative and semi-quantitative invertebrate samples according to standardized procedures. These samples are processed (elutriated and subsampled) in the field to produce as many as four sample components: large-rare, main-body, elutriate, and split. Each sample component is preserved in 10-percent formalin, and two components, large-rare and main-body, are sent to contract laboratories for further processing. The large-rare component is composed of large invertebrates that are removed from the sample matrix during field processing and placed in one or more containers. The main-body sample component consists of the remaining sample materials (sediment, detritus, and invertebrates) and is subsampled in the field to achieve a volume of 750 milliliters or less. The remaining two sample components, elutriate and split, are used for quality-assurance and quality-control purposes. Contract laboratories are used to identify and quantify invertebrates from the large-rare and main-body sample components according to the procedures and guidelines specified within this document. These guidelines allow the use of subsampling techniques to reduce the volume of sample material processed and to facilitate identifications. These processing procedures and techniques may be modified if the modifications provide equal or greater levels of accuracy and precision. The intent of sample processing is to determine the quantity of each taxon present in the semi-quantitative samples or to list the taxa present in qualitative samples. The processing guidelines provide standardized laboratory forms, sample labels, detailed sample processing flow charts, standardized format for electronic data, quality-assurance procedures and checks, sample tracking standards, and target levels for taxonomic determinations. The contract laboratory (1) is responsible for identifications and quantifications, (2) constructs reference collections, (3) provides data in hard copy and electronic forms, (4) follows specified quality-assurance and quality-control procedures, and (5) returns all processed and unprocessed portions of the samples. The U.S. Geological Survey's Quality Management Group maintains a Biological Quality-Assurance Unit, located at the National Water-Quality Laboratory, Arvada, Colorado, to oversee the use of contract laboratories and ensure the quality of data obtained from these laboratories according to the guidelines established in this document. This unit establishes contract specifications, reviews contractor performance (timeliness, accuracy, and consistency), enters data into the National Water Information System-II data base, maintains in-house reference collections, deposits voucher specimens in outside museums, and interacts with taxonomic experts within and outside the U.S. Geological Survey. This unit also modifies the existing sample processing and quality-assurance guidelines, establishes criteria and testing procedures for qualifying potential contract laboratories, identifies qualified taxonomic experts, and establishes voucher collections.
Automation in clinical bacteriology: what system to choose?
Greub, G; Prod'hom, G
2011-05-01
With increased activity and reduced financial and human resources, there is a need for automation in clinical bacteriology. Initial processing of clinical samples includes repetitive and fastidious steps. These tasks are suitable for automation, and several instruments are now available on the market, including the WASP (Copan), Previ-Isola (BioMerieux), Innova (Becton-Dickinson) and Inoqula (KIESTRA) systems. These new instruments allow efficient and accurate inoculation of samples, including four main steps: (i) selecting the appropriate Petri dish; (ii) inoculating the sample; (iii) spreading the inoculum on agar plates to obtain, upon incubation, well-separated bacterial colonies; and (iv) accurate labelling and sorting of each inoculated media. The challenge for clinical bacteriologists is to determine what is the ideal automated system for their own laboratory. Indeed, different solutions will be preferred, according to the number and variety of samples, and to the types of sample that will be processed with the automated system. The final choice is troublesome, because audits proposed by industrials risk being biased towards the solution proposed by their company, and because these automated systems may not be easily tested on site prior to the final decision, owing to the complexity of computer connections between the laboratory information system and the instrument. This article thus summarizes the main parameters that need to be taken into account for choosing the optimal system, and provides some clues to help clinical bacteriologists to make their choice. © 2011 The Authors. Clinical Microbiology and Infection © 2011 European Society of Clinical Microbiology and Infectious Diseases.
Comparison of human umbilical cord blood processing with or without hydroxyethyl starch.
Souri, Milad; Nikougoftar Zarif, Mahin; Rasouli, Mahboobeh; Golzadeh, Khadijeh; Nakhlestani Hagh, Mozhdeh; Ezzati, Nasim; Atarodi, Kamran
2017-11-01
Umbilical cord blood (UCB) processing with hydroxyethyl starch (HES) is the most common protocol in the cord blood banks. The quality of UCB volume reduction was guaranteed by minimum manipulation of cord blood samples in the closed system. This study aimed to analyze and compare cell recovery and viability of UCB processed using the Sepax automated system in the presence and absence of HES. Thirty UCB bags with a total nucleated cell (TNC) count of more than 2.5 × 10 9 were divided in two bags with equal volume. HES solution was added to one bag and another was intact. Both bags were processed with the Sepax. To determine cell recovery, viability, and potential of colony-forming cells (CFCs), preprocessing, postprocessing, and thawing samples were analyzed. The mean TNC recovery after processing and after thaw was significantly better with the HES method (p < 0.01 for the postprocessing step and p < 0.05 for the postthaw step). There were no significant differences to mononucleated cells (MNCs) and CD34+ cell recovery between the two methods after processing and after thaw. TNC and MNC viability was significantly higher without HES after processing and after thaw (p < 0.01). The results of the CFC assay were similar for both methods after processing and after thaw. These results showed that processing of UCB using the Sepax system with the without-HES protocol due to the lower manipulation of samples could be used as an eligible protocol to reduce the volume of UCB. © 2017 AABB.
MSL Chemistry and Mineralogy X-Ray Diffraction X-Ray Fluorescence (CheMin) Instrument
NASA Technical Reports Server (NTRS)
Zimmerman, Wayne; Blake, Dave; Harris, William; Morookian, John Michael; Randall, Dave; Reder, Leonard J.; Sarrazin, Phillipe
2013-01-01
This paper provides an overview of the Mars Science Laboratory (MSL) Chemistry and Mineralogy Xray Diffraction (XRD), X-ray Fluorescence (XRF) (CheMin) Instrument, an element of the landed Curiosity rover payload, which landed on Mars in August of 2012. The scientific goal of the MSL mission is to explore and quantitatively assess regions in Gale Crater as a potential habitat for life - past or present. The CheMin instrument will receive Martian rock and soil samples from the MSL Sample Acquisition/Sample Processing and Handling (SA/SPaH) system, and process it utilizing X-Ray spectroscopy methods to determine mineral composition. The Chemin instrument will analyze Martian soil and rocks to enable scientists to investigate geophysical processes occurring on Mars. The CheMin science objectives and proposed surface operations are described along with the CheMin hardware with an emphasis on the system engineering challenges associated with developing such a complex instrument.
Continuous flow electrophoresis system experiments on shuttle flights STS-6 and STS-7
NASA Technical Reports Server (NTRS)
Snyder, Robert S.; Rhodes, Percy H.; Miller, Teresa Y.
1988-01-01
The development of a space continuous flow electrophoresis system (CFES) is discussed. The objectives of the experiment were: (1) to use a model sample material at a high concentration to evaluate the continuous flow electrophoresis process in the McDonnell Douglass CFES instrument and compare its separation resolution and sample throughput with related devices on Earth, and (2) to expand the basic knowledge of the limitations imposed by fluid flows and particle concentration effects on the electrophoresis process by careful design and evaluation of the space experiment. Hemoglobin and polysaccharide were selected as samples of concentration effects. The results from space show a large band spread of the high concentration of the single species of hemoglobin that was principally due to the mismatch of electrical conductivity between the sample and buffer.
Defining And Characterizing Sample Representativeness For DWPF Melter Feed Samples
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shine, E. P.; Poirier, M. R.
2013-10-29
Representative sampling is important throughout the Defense Waste Processing Facility (DWPF) process, and the demonstrated success of the DWPF process to achieve glass product quality over the past two decades is a direct result of the quality of information obtained from the process. The objective of this report was to present sampling methods that the Savannah River Site (SRS) used to qualify waste being dispositioned at the DWPF. The goal was to emphasize the methodology, not a list of outcomes from those studies. This methodology includes proven methods for taking representative samples, the use of controlled analytical methods, and datamore » interpretation and reporting that considers the uncertainty of all error sources. Numerous sampling studies were conducted during the development of the DWPF process and still continue to be performed in order to evaluate options for process improvement. Study designs were based on use of statistical tools applicable to the determination of uncertainties associated with the data needs. Successful designs are apt to be repeated, so this report chose only to include prototypic case studies that typify the characteristics of frequently used designs. Case studies have been presented for studying in-tank homogeneity, evaluating the suitability of sampler systems, determining factors that affect mixing and sampling, comparing the final waste glass product chemical composition and durability to that of the glass pour stream sample and other samples from process vessels, and assessing the uniformity of the chemical composition in the waste glass product. Many of these studies efficiently addressed more than one of these areas of concern associated with demonstrating sample representativeness and provide examples of statistical tools in use for DWPF. The time when many of these designs were implemented was in an age when the sampling ideas of Pierre Gy were not as widespread as they are today. Nonetheless, the engineers and statisticians used carefully thought out designs that systematically and economically provided plans for data collection from the DWPF process. Key shared features of the sampling designs used at DWPF and the Gy sampling methodology were the specification of a standard for sample representativeness, an investigation that produced data from the process to study the sampling function, and a decision framework used to assess whether the specification was met based on the data. Without going into detail with regard to the seven errors identified by Pierre Gy, as excellent summaries are readily available such as Pitard [1989] and Smith [2001], SRS engineers understood, for example, that samplers can be biased (Gy's extraction error), and developed plans to mitigate those biases. Experiments that compared installed samplers with more representative samples obtained directly from the tank may not have resulted in systematically partitioning sampling errors into the now well-known error categories of Gy, but did provide overall information on the suitability of sampling systems. Most of the designs in this report are related to the DWPF vessels, not the large SRS Tank Farm tanks. Samples from the DWPF Slurry Mix Evaporator (SME), which contains the feed to the DWPF melter, are characterized using standardized analytical methods with known uncertainty. The analytical error is combined with the established error from sampling and processing in DWPF to determine the melter feed composition. This composition is used with the known uncertainty of the models in the Product Composition Control System (PCCS) to ensure that the wasteform that is produced is comfortably within the acceptable processing and product performance region. Having the advantage of many years of processing that meets the waste glass product acceptance criteria, the DWPF process has provided a considerable amount of data about itself in addition to the data from many special studies. Demonstrating representative sampling directly from the large Tank Farm tanks is a difficult, if not unsolvable enterprise due to limited accessibility. However, the consistency and the adequacy of sampling and mixing at SRS could at least be studied under the controlled process conditions based on samples discussed by Ray and others [2012a] in Waste Form Qualification Report (WQR) Volume 2 and the transfers from Tanks 40H and 51H to the Sludge Receipt and Adjustment Tank (SRAT) within DWPF. It is important to realize that the need for sample representativeness becomes more stringent as the material gets closer to the melter, and the tanks within DWPF have been studied extensively to meet those needs.« less
Massively parallel nanowell-based single-cell gene expression profiling.
Goldstein, Leonard D; Chen, Ying-Jiun Jasmine; Dunne, Jude; Mir, Alain; Hubschle, Hermann; Guillory, Joseph; Yuan, Wenlin; Zhang, Jingli; Stinson, Jeremy; Jaiswal, Bijay; Pahuja, Kanika Bajaj; Mann, Ishminder; Schaal, Thomas; Chan, Leo; Anandakrishnan, Sangeetha; Lin, Chun-Wah; Espinoza, Patricio; Husain, Syed; Shapiro, Harris; Swaminathan, Karthikeyan; Wei, Sherry; Srinivasan, Maithreyan; Seshagiri, Somasekar; Modrusan, Zora
2017-07-07
Technological advances have enabled transcriptome characterization of cell types at the single-cell level providing new biological insights. New methods that enable simple yet high-throughput single-cell expression profiling are highly desirable. Here we report a novel nanowell-based single-cell RNA sequencing system, ICELL8, which enables processing of thousands of cells per sample. The system employs a 5,184-nanowell-containing microchip to capture ~1,300 single cells and process them. Each nanowell contains preprinted oligonucleotides encoding poly-d(T), a unique well barcode, and a unique molecular identifier. The ICELL8 system uses imaging software to identify nanowells containing viable single cells and only wells with single cells are processed into sequencing libraries. Here, we report the performance and utility of ICELL8 using samples of increasing complexity from cultured cells to mouse solid tissue samples. Our assessment of the system to discriminate between mixed human and mouse cells showed that ICELL8 has a low cell multiplet rate (< 3%) and low cross-cell contamination. We characterized single-cell transcriptomes of more than a thousand cultured human and mouse cells as well as 468 mouse pancreatic islets cells. We were able to identify distinct cell types in pancreatic islets, including alpha, beta, delta and gamma cells. Overall, ICELL8 provides efficient and cost-effective single-cell expression profiling of thousands of cells, allowing researchers to decipher single-cell transcriptomes within complex biological samples.
Designing to Sample the Unknown: Lessons from OSIRIS-REx Project Systems Engineering
NASA Technical Reports Server (NTRS)
Everett, David; Mink, Ronald; Linn, Timothy; Wood, Joshua
2017-01-01
On September 8, 2016, the third NASA New Frontiers mission launched on an Atlas V 411. The Origins, Spectral Interpretation, Resource Identification, Security-Regolith Explorer (OSIRIS-REx) will rendezvous with asteroid Bennu in 2018, collect a sample in 2020, and return that sample to Earth in September 2023. The development team has overcome a number of challenges in order to design and build a system that will make contact with an unexplored, airless, low-gravity body. This paper will provide an overview of the mission, then focus in on the system-level challenges and some of the key system-level processes. Some of the lessons here are unique to the type of mission, like discussion of operating at a largely-unknown, low-gravity object. Other lessons, particularly from the build phase, have broad implications. The OSIRIS-REx risk management process was particularly effective in achieving an on-time and under-budget development effort. The systematic requirements management and verification and the system validation also helped identify numerous potential problems. The final assessment of the OSIRIS-REx performance will need to wait until the sample is returned in 2023, but this post-launch assessment will capture some of the key systems-engineering lessons from the development team.
[Quality Management System in Pathological Laboratory].
Koyatsu, Junichi; Ueda, Yoshihiko
2015-07-01
Even compared to other clinical laboratories, the pathological laboratory conducts troublesome work, and many of the work processes are also manual. Therefore, the introduction of the systematic management of administration is necessary. It will be a shortcut to use existing standards such as ISO 15189 for this purpose. There is no standard specialized for the pathological laboratory, but it is considered to be important to a pathological laboratory in particular. 1. Safety nianagement of the personnel and environmental conditions. Comply with laws and regulations concerning the handling of hazardous materials. 2. Pre-examination processes. The laboratory shall have documented procedures for the proper collection and handling of primary samples. Developed and documented criteria for acceptance or rejection of samples are applied. 3. Examination processes. Selection, verification, and validation of the examination procedures. Devise a system that can constantly monitor the traceability of the sample. 4. Post-examination processes. Storage, retention, and disposal of clinical samples. 5. Release of results. When examination results fall within established alert or critical intervals, immediately notify the physicians. The important point is to recognize the needs of the client and be aware that pathological diagnoses are always "the final diagnoses".
Data processing for water monitoring system
NASA Technical Reports Server (NTRS)
Monford, L.; Linton, A. T.
1978-01-01
Water monitoring data acquisition system is structured about central computer that controls sampling and sensor operation, and analyzes and displays data in real time. Unit is essentially separated into two systems: computer system, and hard wire backup system which may function separately or with computer.
NASA Technical Reports Server (NTRS)
Hudson, Nicolas; Lin, Ying; Barengoltz, Jack
2010-01-01
A method for evaluating the probability of a Viable Earth Microorganism (VEM) contaminating a sample during the sample acquisition and handling (SAH) process of a potential future Mars Sample Return mission is developed. A scenario where multiple core samples would be acquired using a rotary percussive coring tool, deployed from an arm on a MER class rover is analyzed. The analysis is conducted in a structured way by decomposing sample acquisition and handling process into a series of discrete time steps, and breaking the physical system into a set of relevant components. At each discrete time step, two key functions are defined: The probability of a VEM being released from each component, and the transport matrix, which represents the probability of VEM transport from one component to another. By defining the expected the number of VEMs on each component at the start of the sampling process, these decompositions allow the expected number of VEMs on each component at each sampling step to be represented as a Markov chain. This formalism provides a rigorous mathematical framework in which to analyze the probability of a VEM entering the sample chain, as well as making the analysis tractable by breaking the process down into small analyzable steps.
NASA Astrophysics Data System (ADS)
Wang, Zhong-Jie; Ni, Wen; Li, Ke-Qing; Huang, Xiao-Yan; Zhu, Li-Ping
2011-08-01
The crystallization process of iron-rich glass-ceramics prepared from the mixture of nickel slag (NS) and blast furnace slag (BFS) with a small amount of quartz sand was investigated. A modified melting method which was more energy-saving than the traditional methods was used to control the crystallization process. The results show that the iron-rich system has much lower melting temperature, glass transition temperature ( T g), and glass crystallization temperature ( T c), which can result in a further energy-saving process. The results also show that the system has a quick but controllable crystallization process with its peak crystallization temperature at 918°C. The crystallization of augite crystals begins from the edge of the sample and invades into the whole sample. The crystallization process can be completed in a few minutes. A distinct boundary between the crystallized part and the non-crystallized part exists during the process. In the non-crystallized part showing a black colour, some sphere-shaped augite crystals already exist in the glass matrix before samples are heated to T c. In the crystallized part showing a khaki colour, a compact structure is formed by augite crystals.
System for high throughput water extraction from soil material for stable isotope analysis of water
USDA-ARS?s Scientific Manuscript database
A major limitation in the use of stable isotope of water in ecological studies is the time that is required to extract water from soil and plant samples. Using vacuum distillation the extraction time can be less than one hour per sample. Therefore, assembling a distillation system that can process m...
Automated high-throughput flow-through real-time diagnostic system
Regan, John Frederick
2012-10-30
An automated real-time flow-through system capable of processing multiple samples in an asynchronous, simultaneous, and parallel fashion for nucleic acid extraction and purification, followed by assay assembly, genetic amplification, multiplex detection, analysis, and decontamination. The system is able to hold and access an unlimited number of fluorescent reagents that may be used to screen samples for the presence of specific sequences. The apparatus works by associating extracted and purified sample with a series of reagent plugs that have been formed in a flow channel and delivered to a flow-through real-time amplification detector that has a multiplicity of optical windows, to which the sample-reagent plugs are placed in an operative position. The diagnostic apparatus includes sample multi-position valves, a master sample multi-position valve, a master reagent multi-position valve, reagent multi-position valves, and an optical amplification/detection system.
The release of persistent organic pollutants from a closed system dicofol production process.
Li, Sumei; Tian, Yajing; Ding, Qiong; Liu, Wenbin
2014-01-01
High concentrations of polychlorinated dibenzo-p-dioxins and dibenzofurans (PCDD/Fs) have been found to be produced in chemical processes in which chlorine is a raw material. Samples of workshop air, waste water, waste acid, and the dicofol product were collected from a pesticide factory in China that uses a closed-system dicofol production process, and were analyzed for PCDD/Fs and ΣDDTs. The ΣDDTs concentrations were 1.88-17.53 μg m(-3) in the workshop air samples, 4.85-456 μg kg(-1) in the waste water and waste acid samples, and 4.74 g kg(-1) in the dicofol product. The total estimated daily intakes of ΣDDTs for workers by inhalation in the workplace were in the range of 0.38-3.51 μg kg(-1)bwd(-1) for moderate activities. The annual amounts of ΣDDTs and p,p'-DDT directly released to the environment via the use of dicofol were 9,480 kg and 1,080 kg, respectively. The PCDD/F toxicity equivalent values (I-TEQs) in the waste water and waste acid samples ranged from 1.5 to 122 pg I-TEQ kg(-1) and 86.3 ng I-TEQ kg(-1) in the dicofol sample. The annual amount of PCDD/Fs released to the environment was 0.17 g I-TEQ. From the PCDD/F distribution patterns, it is suggested that the major pathway for PCDD/F formation involves precursor synthesis during the production of dicofol in the closed-system process. Copyright © 2013 Elsevier Ltd. All rights reserved.
Field Immune Assessment during Simulated Planetary Exploration in the Canadian Arctic
NASA Technical Reports Server (NTRS)
Crucian, Brian; Lee, Pascal; Stowe, Raymond; Jones, Jeff; Effenhauser, Rainer; Widen, Raymond; Sams, Clarence
2006-01-01
Dysregulation of the immune system has been shown to occur during space flight, although the detailed nature of the phenomenon and the clinical risks for exploration class missions has yet to be established. In addition, the growing clinical significance of immune system evaluation combined with epidemic infectious disease rates in third world countries provides a strong rationale for the development of field-compatible clinical immunology techniques and equipment. In July 2002 NASA performed a comprehensive field immunology assessment on crewmembers participating in the Haughton-Mars Project (HMP) on Devon Island in the high Canadian Arctic. The purpose of the study was to evaluate mission-associated effects on the human immune system, as well as to evaluate techniques developed for processing immune samples in remote field locations. Ten HMP-2002 participants volunteered for the study. A field protocol was developed at NASA-JSC for performing sample collection, blood staining/processing for immunophenotype analysis, wholeblood mitogenic culture for functional assessments and cell-sample preservation on-location at Devon Island. Specific assays included peripheral leukocyte distribution; constitutively activated T cells, intracellular cytokine profiles and plasma EBV viral antibody levels. Study timepoints were L-30, midmission and R+60. The protocol developed for immune sample processing in remote field locations functioned properly. Samples were processed in the field location, and stabilized for subsequent analysis at the Johnson Space Center in Houston. The data indicated that some phenotype, immune function and stress hormone changes occurred in the HMP field participants that were largely distinct from pre-mission baseline and post-mission recovery data. These immune changes appear similar to those observed in Astronauts following spaceflight. The sample processing protocol developed for this study may have applications for immune assessment during exploration-class space missions or in remote terrestrial field locations. The data validate the use of the HMP as a ground-based spaceflight/planetary exploration analog for some aspects of human physiology.
ISOLOK VALVE ACCEPTANCE TESTING FOR DWPF SME SAMPLING PROCESS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Edwards, T.; Hera, K.; Coleman, C.
2011-12-05
Evaluation of the Defense Waste Processing Facility (DWPF) Chemical Process Cell (CPC) cycle time identified several opportunities to improve the CPC processing time. Of the opportunities, a focus area related to optimizing the equipment and efficiency of the sample turnaround time for DWPF Analytical Laboratory was identified. The Mechanical Systems & Custom Equipment Development (MS&CED) Section of the Savannah River National Laboratory (SRNL) evaluated the possibility of using an Isolok{reg_sign} sampling valve as an alternative to the Hydragard{reg_sign} valve for taking process samples. Previous viability testing was conducted with favorable results using the Isolok sampler and reported in SRNL-STI-2010-00749 (1).more » This task has the potential to improve operability, reduce maintenance time and decrease CPC cycle time. This report summarizes the results from acceptance testing which was requested in Task Technical Request (TTR) HLW-DWPF-TTR-2010-0036 (2) and which was conducted as outlined in Task Technical and Quality Assurance Plan (TTQAP) SRNL-RP-2011-00145 (3). The Isolok to be tested is the same model which was tested, qualified, and installed in the Sludge Receipt Adjustment Tank (SRAT) sample system. RW-0333P QA requirements apply to this task. This task was to qualify the Isolok sampler for use in the DWPF Slurry Mix Evaporator (SME) sampling process. The Hydragard, which is the current baseline sampling method, was used for comparison to the Isolok sampling data. The Isolok sampler is an air powered grab sampler used to 'pull' a sample volume from a process line. The operation of the sampler is shown in Figure 1. The image on the left shows the Isolok's spool extended into the process line and the image on the right shows the sampler retracted and then dispensing the liquid into the sampling container. To determine tank homogeneity, a Coliwasa sampler was used to grab samples at a high and low location within the mixing tank. Data from the two locations were compared to determine if the contents of the tank were well mixed. The Coliwasa sampler is a tube with a stopper at the bottom and is designed to obtain grab samples from specific locations within the drum contents. A position paper (4) was issued to address the prototypic flow loop issues and simulant selections. A statistically designed plan (5) was issued to address the total number of samples each sampler needed to pull, to provide the random order in which samples were pulled and to group samples for elemental analysis. The TTR required that the Isolok sampler perform as well as the Hydragard sampler during these tests to ensure the acceptability of the Isolok sampler for use in the DWPF sampling cells. Procedure No.L9.4-5015 was used to document the sample parameters and process steps. Completed procedures are located in R&D Engineering job folder 23269.« less
Automated high-throughput protein purification using an ÄKTApurifier and a CETAC autosampler.
Yoo, Daniel; Provchy, Justin; Park, Cynthia; Schulz, Craig; Walker, Kenneth
2014-05-30
As the pace of drug discovery accelerates there is an increased focus on screening larger numbers of protein therapeutic candidates to identify those that are functionally superior and to assess manufacturability earlier in the process. Although there have been advances toward high throughput (HT) cloning and expression, protein purification is still an area where improvements can be made to conventional techniques. Current methodologies for purification often involve a tradeoff between HT automation or capacity and quality. We present an ÄKTA combined with an autosampler, the ÄKTA-AS, which has the capability of purifying up to 240 samples in two chromatographic dimensions without the need for user intervention. The ÄKTA-AS has been shown to be reliable with sample volumes between 0.5 mL and 100 mL, and the innovative use of a uniquely configured loading valve ensures reliability by efficiently removing air from the system as well as preventing sample cross contamination. Incorporation of a sample pump flush minimizes sample loss and enables recoveries ranging from the low tens of micrograms to milligram quantities of protein. In addition, when used in an affinity capture-buffer exchange format the final samples are formulated in a buffer compatible with most assays without requirement of additional downstream processing. The system is designed to capture samples in 96-well microplate format allowing for seamless integration of downstream HT analytic processes such as microfluidic or HPLC analysis. Most notably, there is minimal operator intervention to operate this system, thereby increasing efficiency, sample consistency and reducing the risk of human error. Copyright © 2014 Elsevier B.V. All rights reserved.
NASA Technical Reports Server (NTRS)
Voellmer, George
1997-01-01
The Goddard Space Flight Center has developed the Robot Operated Materials Processing System (ROMPS) that flew aboard STS-64 in September, 1994. The ROMPS robot transported pallets containing wafers of different materials from their storage racks to a furnace for thermal processing. A system of tapered guides and compliant springs was designed to deal with the potential misalignments. The robot and all the sample pallets were locked down for launch and landing. The design of the passive lockdown system, and the interplay between it and the alignment system are presented.
On the enhanced sampling over energy barriers in molecular dynamics simulations.
Gao, Yi Qin; Yang, Lijiang
2006-09-21
We present here calculations of free energies of multidimensional systems using an efficient sampling method. The method uses a transformed potential energy surface, which allows an efficient sampling of both low and high energy spaces and accelerates transitions over barriers. It allows efficient sampling of the configuration space over and only over the desired energy range(s). It does not require predetermined or selected reaction coordinate(s). We apply this method to study the dynamics of slow barrier crossing processes in a disaccharide and a dipeptide system.
Demonstration of laser speckle system on burner liner cyclic rig
NASA Technical Reports Server (NTRS)
Stetson, K. A.
1986-01-01
A demonstration test was conducted to apply speckle photogrammetry to the measurement of strains on a sample of combustor liner material in a cyclic fatigue rig. A system for recording specklegrams was assembled and shipped to the NASA Lewis Research Center, where it was set up and operated during rig tests. Data in the form of recorded specklegrams were sent back to United Technologies Research Center for processing to extract strains. Difficulties were found in the form of warping and bowing of the sample during the tests which degraded the data. Steps were taken by NASA personnel to correct this problem and further tests were run. Final data processing indicated erratic patterns of strain on the burner liner sample.
Real time viability detection of bacterial spores
Vanderberg, Laura A.; Herdendorf, Timothy J.; Obiso, Richard J.
2003-07-29
This invention relates to a process for detecting the presence of viable bacterial spores in a sample and to a spore detection system, the process including placing a sample in a germination medium for a period of time sufficient for commitment of any present viable bacterial spores to occur, mixing the sample with a solution of a lanthanide capable of forming a fluorescent complex with dipicolinic acid, and, measuring the sample for the presence of dipicolinic acid, and the system including a germination chamber having inlets from a sample chamber, a germinant chamber and a bleach chamber, the germination chamber further including an outlet through a filtering means, the outlet connected to a detection chamber, the detection chamber having an inlet from a fluorescence promoting metal chamber and the detection chamber including a spectral excitation source and a means of measuring emission spectra from a sample, the detection chamber further connected to a waste chamber. A germination reaction mixture useful for promoting commitment of any viable bacterial spores in a sample including a combination of L-alanine, L-asparagine and D-glucose is also described.
The Sample Analysis at Mars Investigation and Instrument Suite
NASA Technical Reports Server (NTRS)
Mahaffy, Paul; Webster, Christopher R.; Conrad, Pamela G.; Arvey, Robert; Bleacher, Lora; Brinckerhoff, William B.; Eigenbrode, Jennifer L.; Chalmers, Robert A.; Dworkin, Jason P.; Errigo, Therese;
2012-01-01
The Sample Analysis at Mars (SAM) investigation of the Mars Science Laboratory (MSL) addresses the chemical and isotopic composition of the atmosphere and volatiles extracted from solid samples. The SAM investigation is designed to contribute substantially to the mission goal of quantitatively assessing the habitability of Mars as an essential step in the search for past or present life on Mars. SAM is a 40 kg instrument suite located in the interior of MSL's Curiosity rover. The SAM instruments are a quadrupole mass spectrometer, a tunable laser spectrometer, and a 6-column gas chromatograph all coupled through solid and gas processing systems to provide complementary information on the same samples. The SAM suite is able to measure a suite of light isotopes and to analyze volatiles directly from the atmosphere or thermally released from solid samples. In addition to measurements of simple inorganic compounds and noble gases SAM will conduct a sensitive search for organic compounds with either thermal or chemical extraction from sieved samples delivered by the sample processing system on the Curiosity rover's robotic arm,
Where do the Field Plots Belong? A Multiple-Constraint Sampling Design for the BigFoot Project
NASA Astrophysics Data System (ADS)
Kennedy, R. E.; Cohen, W. B.; Kirschbaum, A. A.; Gower, S. T.
2002-12-01
A key component of a MODIS validation project is effective characterization of biophysical measures on the ground. Fine-grain ecological field measurements must be placed strategically to capture variability at the scale of the MODIS imagery. Here we describe the BigFoot project's revised sampling scheme, designed to simultaneously meet three important goals: capture landscape variability, avoid spatial autocorrelation between field plots, and minimize time and expense of field sampling. A stochastic process places plots in clumped constellations to reduce field sampling costs, while minimizing spatial autocorrelation. This stochastic process is repeated, creating several hundred realizations of plot constellations. Each constellation is scored and ranked according to its ability to match landscape variability in several Landsat-based spectral indices, and its ability to minimize field sampling costs. We show how this approach has recently been used to place sample plots at the BigFoot project's two newest study areas, one in a desert system and one in a tundra system. We also contrast this sampling approach to that already used at the four prior BigFoot project sites.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Weber, J. K. R.; Alderman, O. L. G.; Advanced Photon Source, Argonne National Laboratory, Argonne, Illinois 60439
2016-07-15
An aerodynamic levitator with carbon dioxide laser beam heating was integrated with a hermetically sealed controlled atmosphere chamber and sample handling mechanism. The system enabled containment of radioactive samples and control of the process atmosphere chemistry. The chamber was typically operated at a pressure of approximately 0.9 bars to ensure containment of the materials being processed. Samples 2.5-3 mm in diameter were levitated in flowing gas to achieve containerless conditions. Levitated samples were heated to temperatures of up to 3500 °C with a partially focused carbon dioxide laser beam. Sample temperature was measured using an optical pyrometer. The sample environment wasmore » integrated with a high energy (100 keV) x-ray synchrotron beamline to enable in situ structure measurements to be made on levitated samples as they were heated, melted, and supercooled. The system was controlled from outside the x-ray beamline hutch by using a LabVIEW program. Measurements have been made on hot solid and molten uranium dioxide and binary uranium dioxide-zirconium dioxide compositions.« less
NASA Astrophysics Data System (ADS)
Duan, Yixiang; Su, Yongxuan; Jin, Zhe; Abeln, Stephen P.
2000-03-01
The development of a highly sensitive, field portable, low-powered instrument for on-site, real-time liquid waste stream monitoring is described in this article. A series of factors such as system sensitivity and portability, plasma source, sample introduction, desolvation system, power supply, and the instrument configuration, were carefully considered in the design of the portable instrument. A newly designed, miniature, modified microwave plasma source was selected as the emission source for spectroscopy measurement, and an integrated small spectrometer with a charge-coupled device detector was installed for signal processing and detection. An innovative beam collection system with optical fibers was designed and used for emission signal collection. Microwave plasma can be sustained with various gases at relatively low power, and it possesses high detection capabilities for both metal and nonmetal pollutants, making it desirable to use for on-site, real-time, liquid waste stream monitoring. An effective in situ sampling system was coupled with a high efficiency desolvation device for direct-sampling liquid samples into the plasma. A portable computer control system is used for data processing. The new, integrated instrument can be easily used for on-site, real-time monitoring in the field. The system possesses a series of advantages, including high sensitivity for metal and nonmetal elements; in situ sampling; compact structure; low cost; and ease of operation and handling. These advantages will significantly overcome the limitations of previous monitoring techniques and make great contributions to environmental restoration and monitoring.
Fong, Sophia Yui Kau; Poulsen, Jessie; Brandl, Martin; Bauer-Brandl, Annette
2017-01-01
A novel microdialysis-dissolution/permeation (M-D/P) system was developed for the biopharmaceutical assessment of oral drug formulations. This system consists of a side-by-side diffusion chamber, a microdialysis unit fixed within the dissolution chamber for continuous sampling, and a biomimetic Permeapad® as the intestinal barrier. In the M-D/P system, the concentration of the molecularly dissolved drug (with MWCO <20kDa) was measured over time in the dissolution compartment (representing the gastrointestinal tract) while the concentration of the permeated drug was measured in the acceptor compartment (representing the blood). The kinetics of both the dissolution process and the permeation process were simultaneously quantified under circumstances that mimic physiological conditions. For the current proof-of-concept study, hydrocortisone (HCS) in the form of slowly dissolving solvate crystals and buffer and the biorelevant fasted state simulated intestinal fluids (FaSSIF), were employed as the model drug and dissolution media, respectively. The applicability of the M-D/P system to dissolution and permeation profiling of HCS in buffer and in FaSSIF has been successfully demonstrated. Compared to the conventional direct sampling method (using filter of 0.1-0.45μm), sampling by the M-D/P system exhibited distinct advantages, including (1) showing minimal disturbance of the permeation process, (2) differentiating "molecularly" dissolved drugs from "apparently" dissolved drugs during dissolution of HCS in FaSSIF, and (3) being less laborious and having better sampling temporal resolution. M-D/P system appeared to be a promising, simple and routine tool that allows for the researchers' intensive comprehension of the interplay of dissolution and permeation thus helping for better oral formulation screening and as an ultimate goal, for better dosage forms assessment. Copyright © 2016. Published by Elsevier B.V.
Mayberry, Addison; Perkins, David L.; Holcomb, Daniel E.
2018-01-01
Miniaturized imaging devices have pushed the boundaries of point-of-care imaging, but existing mobile-phone-based imaging systems do not exploit the full potential of smart phones. This work demonstrates the use of simple imaging configurations to deliver superior image quality and the ability to handle a wide range of biological samples. Results presented in this work are from analysis of fluorescent beads under fluorescence imaging, as well as helminth eggs and freshwater mussel larvae under white light imaging. To demonstrate versatility of the systems, real time analysis and post-processing results of the sample count and sample size are presented in both still images and videos of flowing samples. PMID:29509786
NASA Astrophysics Data System (ADS)
Yan, Yifang; Yang, Chunyu; Ma, Xiaoping; Zhou, Linna
2018-02-01
In this paper, sampled-data H∞ filtering problem is considered for Markovian jump singularly perturbed systems with time-varying delay and missing measurements. The sampled-data system is represented by a time-delay system, and the missing measurement phenomenon is described by an independent Bernoulli random process. By constructing an ɛ-dependent stochastic Lyapunov-Krasovskii functional, delay-dependent sufficient conditions are derived such that the filter error system satisfies the prescribed H∞ performance for all possible missing measurements. Then, an H∞ filter design method is proposed in terms of linear matrix inequalities. Finally, numerical examples are given to illustrate the feasibility and advantages of the obtained results.
An automated atmospheric sampling system operating on 747 airliners
NASA Technical Reports Server (NTRS)
Perkins, P.; Gustafsson, U. R. C.
1975-01-01
An air sampling system that automatically measures the temporal and spatial distribution of selected particulate and gaseous constituents of the atmosphere has been installed on a number of commercial airliners and is collecting data on commercial air routes covering the world. Measurements of constituents related to aircraft engine emissions and other pollutants are made in the upper troposphere and lower stratosphere (6 to 12 km) in support of the Global Air Sampling Program (GASP). Aircraft operated by different airlines sample air at latitudes from the Arctic to Australia. This system includes specialized instrumentation for measuring carbon monoxide, ozone, water vapor, and particulates, a special air inlet probe for sampling outside air, a computerized automatic control, and a data acquisition system. Air constituents and related flight data are tape recorded in flight for later computer processing on the ground.
Chemistry of spring and well waters on Kilauea Volcano, Hawaii, and vicinity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Janik, C.J.; Nathenson, M.; Scholl, M.A.
1994-12-31
Published and new data for chemical and isotopic samples from wells and springs on Kilauea Volcano and vicinity are presented. These data are used to understand processes that determine the chemistry of dilute meteoric water, mixtures with sea water, and thermal water. Data for well and spring samples of non-thermal water indicate that mixing with sea water and dissolution of rock from weathering are the major processes that determine the composition of dissolved constituents in water. Data from coastal springs demonstrate that there is a large thermal system south of the lower east rift of Kilauea. Samples of thermal watermore » from shallow wells in the lower east rift and vicinity have rather variable chemistry indicating that a number of processes operate in the near surface. Water sampled from the available deep wells is different in composition from the shallow thermal water, indicating that generally there is not a significant component of deep water in the shallow wells. Data for samples from available deep wells show significant gradients in chemistry and steam content of the reservoir fluid. These gradients are interpreted to indicate that the reservoir tapped by the existing wells is an evolving vapor-dominated system.« less
Model Transformation for a System of Systems Dependability Safety Case
NASA Technical Reports Server (NTRS)
Murphy, Judy; Driskell, Steve
2011-01-01
The presentation reviews the dependability and safety effort of NASA's Independent Verification and Validation Facility. Topics include: safety engineering process, applications to non-space environment, Phase I overview, process creation, sample SRM artifact, Phase I end result, Phase II model transformation, fault management, and applying Phase II to individual projects.
The effect of processing on the surface physical stability of amorphous solid dispersions.
Yang, Ziyi; Nollenberger, Kathrin; Albers, Jessica; Moffat, Jonathan; Craig, Duncan; Qi, Sheng
2014-11-01
The focus of this study was to investigate the effect of processing on the surface crystallization of amorphous molecular dispersions and gain insight into the mechanisms underpinning this effect. The model systems, amorphous molecular dispersions of felodipine-EUDRAGIT® E PO, were processed both using spin coating (an ultra-fast solvent evaporation based method) and hot melt extrusion (HME) (a melting based method). Amorphous solid dispersions with drug loadings of 10-90% (w/w) were obtained by both processing methods. Samples were stored under 75% RH/room temperatures for up to 10months. Surface crystallization was observed shortly after preparation for the HME samples with high drug loadings (50-90%). Surface crystallization was characterized by powder X-ray diffraction (PXRD), ATR-FTIR spectroscopy and imaging techniques (SEM, AFM and localized thermal analysis). Spin coated molecular dispersions showed significantly higher surface physical stability than hot melt extruded samples. For both systems, the progress of the surface crystal growth followed zero order kinetics on aging. Drug enrichment at the surfaces of HME samples on aging was observed, which may contribute to surface crystallization of amorphous molecular dispersions. In conclusion it was found the amorphous molecular dispersions prepared by spin coating had a significantly higher surface physical stability than the corresponding HME samples, which may be attributed to the increased process-related apparent drug-polymer solubility and reduced molecular mobility due to the quenching effect caused by the rapid solvent evaporation in spin coating. Copyright © 2014 Elsevier B.V. All rights reserved.
How to select a continuous emission monitoring system
DOE Office of Scientific and Technical Information (OSTI.GOV)
Radigan, M.J.
1994-02-01
Selecting a continuous emission monitoring system (CEMS) involves more than picking an analyzer. Successful CEMS interface sampling and data-management systems to produce accurate, reliable reports required by regulatory agencies. Following objective guidelines removes some of the misery from CEMS shopping. However, prospective CEMS buyers should do their homework and develop well-thought-out, detailed specification for the processes' sampling criteria. Fine tuning the analyzer/data management system can eliminate maintenance costs and keep the facility operating within its permit restrictions.
Jin, Choong Eun; Lee, Tae Yoon; Koo, Bonhan; Choi, Kyung-Chul; Chang, Suhwan; Park, Se Yoon; Kim, Ji Yeun; Kim, Sung-Han; Shin, Yong
2017-07-18
The isolation of nucleic acids in the lab on a chip is crucial to achieve the maximal effectiveness of point-of-care testing for detection in clinical applications. Here, we report on the use of a simple and versatile single-channel microfluidic platform that combines dimethyl pimelimidate (DMP) for nucleic acids (both RNA and DNA) extraction without electricity using a thin-film system. The system is based on the adaption of DMP into nonchaotropic-based nucleic acids and the capture of reagents into a low-cost thin-film platform for use as a microfluidic total analysis system, which can be utilized for sample processing in clinical diagnostics. Moreover, we assessed the use of the DMP system for the extraction of nucleic acids from various samples, including mammalian cells, bacterial cells, and viruses from human disease, and we also confirmed that the quality and quantity of the nucleic acids extracted were sufficient to allow for the robust detection of biomarkers and/or pathogens in downstream analysis. Furthermore, this DMP system does not require any instruments and electricity, and has improved time efficiency, portability, and affordability. Thus, we believe that the DMP system may change the paradigm of sample processing in clinical diagnostics.
Automated paleomagnetic and rock magnetic data acquisition with an in-line horizontal "2G" system
NASA Astrophysics Data System (ADS)
Mullender, Tom A. T.; Frederichs, Thomas; Hilgenfeldt, Christian; de Groot, Lennart V.; Fabian, Karl; Dekkers, Mark J.
2016-09-01
Today's paleomagnetic and magnetic proxy studies involve processing of large sample collections while simultaneously demanding high quality data and high reproducibility. Here we describe a fully automated interface based on a commercial horizontal pass-through "2G" DC-SQUID magnetometer. This system is operational at the universities of Bremen (Germany) and Utrecht (Netherlands) since 1998 and 2006, respectively, while a system is currently being built at NGU Trondheim (Norway). The magnetometers are equipped with "in-line" alternating field (AF) demagnetization, a direct-current bias field coil along the coaxial AF demagnetization coil for the acquisition of anhysteretic remanent magnetization (ARM) and a long pulse-field coil for the acquisition of isothermal remanent magnetization (IRM). Samples are contained in dedicated low magnetization perspex holders that are manipulated by a pneumatic pick-and-place-unit. Upon desire samples can be measured in several positions considerably enhancing data quality in particular for magnetically weak samples. In the Bremen system, the peak of the IRM pulse fields is actively measured which reduces the discrepancy between the set field and the field that is actually applied. Techniques for quantifying and removing gyroremanent overprints and for measuring the viscosity of IRM further extend the range of applications of the system. Typically c. 300 paleomagnetic samples can be AF demagnetized per week (15 levels) in the three-position protocol. The versatility of the system is illustrated by several examples of paleomagnetic and rock magnetic data processing.
FOCIS: A forest classification and inventory system using LANDSAT and digital terrain data
NASA Technical Reports Server (NTRS)
Strahler, A. H.; Franklin, J.; Woodcook, C. E.; Logan, T. L.
1981-01-01
Accurate, cost-effective stratification of forest vegetation and timber inventory is the primary goal of a Forest Classification and Inventory System (FOCIS). Conventional timber stratification using photointerpretation can be time-consuming, costly, and inconsistent from analyst to analyst. FOCIS was designed to overcome these problems by using machine processing techniques to extract and process tonal, textural, and terrain information from registered LANDSAT multispectral and digital terrain data. Comparison of samples from timber strata identified by conventional procedures showed that both have about the same potential to reduce the variance of timber volume estimates over simple random sampling.
Imaging system design and image interpolation based on CMOS image sensor
NASA Astrophysics Data System (ADS)
Li, Yu-feng; Liang, Fei; Guo, Rui
2009-11-01
An image acquisition system is introduced, which consists of a color CMOS image sensor (OV9620), SRAM (CY62148), CPLD (EPM7128AE) and DSP (TMS320VC5509A). The CPLD implements the logic and timing control to the system. SRAM stores the image data, and DSP controls the image acquisition system through the SCCB (Omni Vision Serial Camera Control Bus). The timing sequence of the CMOS image sensor OV9620 is analyzed. The imaging part and the high speed image data memory unit are designed. The hardware and software design of the image acquisition and processing system is given. CMOS digital cameras use color filter arrays to sample different spectral components, such as red, green, and blue. At the location of each pixel only one color sample is taken, and the other colors must be interpolated from neighboring samples. We use the edge-oriented adaptive interpolation algorithm for the edge pixels and bilinear interpolation algorithm for the non-edge pixels to improve the visual quality of the interpolated images. This method can get high processing speed, decrease the computational complexity, and effectively preserve the image edges.
Microbiological performance of a food safety management system in a food service operation.
Lahou, E; Jacxsens, L; Daelman, J; Van Landeghem, F; Uyttendaele, M
2012-04-01
The microbiological performance of a food safety management system in a food service operation was measured using a microbiological assessment scheme as a vertical sampling plan throughout the production process, from raw materials to final product. The assessment scheme can give insight into the microbiological contamination and the variability of a production process and pinpoint bottlenecks in the food safety management system. Three production processes were evaluated: a high-risk sandwich production process (involving raw meat preparation), a medium-risk hot meal production process (starting from undercooked raw materials), and a low-risk hot meal production process (reheating in a bag). Microbial quality parameters, hygiene indicators, and relevant pathogens (Listeria monocytogenes, Salmonella, Bacillus cereus, and Escherichia coli O157) were in accordance with legal criteria and/or microbiological guidelines, suggesting that the food safety management system was effective. High levels of total aerobic bacteria (>3.9 log CFU/50 cm(2)) were noted occasionally on gloves of food handlers and on food contact surfaces, especially in high contamination areas (e.g., during handling of raw material, preparation room). Core control activities such as hand hygiene of personnel and cleaning and disinfection (especially in highly contaminated areas) were considered points of attention. The present sampling plan was used to produce an overall microbiological profile (snapshot) to validate the food safety management system in place.
Robotics-assisted mass spectrometry assay platform enabled by open-source electronics.
Chiu, Shih-Hao; Urban, Pawel L
2015-02-15
Mass spectrometry (MS) is an important analytical technique with numerous applications in clinical analysis, biochemistry, environmental analysis, geology and physics. Its success builds on the ability of MS to determine molecular weights of analytes, and elucidate their structures. However, sample handling prior to MS requires a lot of attention and labor. In this work we were aiming to automate processing samples for MS so that analyses could be conducted without much supervision of experienced analysts. The goal of this study was to develop a robotics and information technology-oriented platform that could control the whole analysis process including sample delivery, reaction-based assay, data acquisition, and interaction with the analyst. The proposed platform incorporates a robotic arm for handling sample vials delivered to the laboratory, and several auxiliary devices which facilitate and secure the analysis process. They include: multi-relay board, infrared sensors, photo-interrupters, gyroscopes, force sensors, fingerprint scanner, barcode scanner, touch screen panel, and internet interface. The control of all the building blocks is achieved through implementation of open-source electronics (Arduino), and enabled by custom-written programs in C language. The advantages of the proposed system include: low cost, simplicity, small size, as well as facile automation of sample delivery and processing without the intervention of the analyst. It is envisaged that this simple robotic system may be the forerunner of automated laboratories dedicated to mass spectrometric analysis of biological samples. Copyright © 2014 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Steinbach, Julia; Holmstrand, Henry; Semiletov, Igor; Shakhova, Natalia; Shcherbakova, Kseniia; Kosmach, Denis; Sapart, Célia J.; Gustafsson, Örjan
2015-04-01
We present a method for measurements of the stable and radiocarbon isotope systems of methane in seawater and sediments. The triple isotope characterization of methane is useful in distinguishing different sources and for improving our understanding of biogeochemical processes affecting methane in the water column. D14C-CH4 is an especially powerful addition to stable isotope analyses in distinguishing between thermogenic and biogenic origins of the methane. Such measurements require large sample sizes, due to low natural abundance of the radiocarbon in CH4. Our system for sample collection, methane extraction and purification builds on the approach by Kessler and Reeburgh (Limn. & Ocean. Meth., 2005). An in-field system extracts methane from 30 -120 l water or 1-2 l sediment (depending on the in-situ methane concentration) by purging the samples with Helium to transfer the dissolved methane to the headspace and circulating it through cryogenically cooled absorbent traps where methane is collected. The in-field preparation eliminates the risks of storage and transport of large seawater quantities and subsequent leakage of sample gas as well as ongoing microbial processes and chemical reactions that may alter the sample composition. In the subsequent shore-based treatment, a laboratory system is used to purify and combust the collected CH4 to AMS-amenable CO2. Subsamples from the methane traps are analyzed for stable isotopes and compared to stable isotope measurements directly measured from small water samples taken in parallel, to correct for any potential fractionation occurring during this process. The system has been successfully tested and used on several shorter shipboard expeditions in the Baltic Sea and on a long summer expedition across the Arctic Ocean. Here we present the details of the method and testing, as well as first triple isotope field data from two cruises to the Landsort Deep area in the Central Baltic Sea.
Study of sample drilling techniques for Mars sample return missions
NASA Technical Reports Server (NTRS)
Mitchell, D. C.; Harris, P. T.
1980-01-01
To demonstrate the feasibility of acquiring various surface samples for a Mars sample return mission the following tasks were performed: (1) design of a Mars rover-mounted drill system capable of acquiring crystalline rock cores; prediction of performance, mass, and power requirements for various size systems, and the generation of engineering drawings; (2) performance of simulated permafrost coring tests using a residual Apollo lunar surface drill, (3) design of a rock breaker system which can be used to produce small samples of rock chips from rocks which are too large to return to Earth, but too small to be cored with the Rover-mounted drill; (4)design of sample containers for the selected regolith cores, rock cores, and small particulate or rock samples; and (5) design of sample handling and transfer techniques which will be required through all phase of sample acquisition, processing, and stowage on-board the Earth return vehicle. A preliminary design of a light-weight Rover-mounted sampling scoop was also developed.
Monte Carlo sampling in diffusive dynamical systems
NASA Astrophysics Data System (ADS)
Tapias, Diego; Sanders, David P.; Altmann, Eduardo G.
2018-05-01
We introduce a Monte Carlo algorithm to efficiently compute transport properties of chaotic dynamical systems. Our method exploits the importance sampling technique that favors trajectories in the tail of the distribution of displacements, where deviations from a diffusive process are most prominent. We search for initial conditions using a proposal that correlates states in the Markov chain constructed via a Metropolis-Hastings algorithm. We show that our method outperforms the direct sampling method and also Metropolis-Hastings methods with alternative proposals. We test our general method through numerical simulations in 1D (box-map) and 2D (Lorentz gas) systems.
Al-Azmi, D; Snopek, B; Sayed, A M; Domanski, T
2004-01-01
Based on the different levels of solubility of radon gas in organic solvents and water, a bubbling system has been developed to transfer radon gas, dissolving naturally in water samples, to an organic solvent, i.e. olive oil, which is known to be a good solvent of radon gas. The system features the application of a fixed volume of bubbling air by introducing a fixed volume of water into a flask mounted above the system, to displace an identical volume of air from an air cylinder. Thus a gravitational flow of water is provided without the need for pumping. Then, the flushing air (radon-enriched air) is directed through a vial containing olive oil, to achieve deposition of the radon gas by another bubbling process. Following this, the vial (containing olive oil) is measured by direct use of gamma ray spectrometry, without the need of any chemical or physical processing of the samples. Using a standard solution of 226Ra/222Rn, a lowest measurable concentration (LMC) of radon in water samples of 9.4 Bq L(-1) has been achieved (below the maximum contaminant level of 11 Bq L(-1)).
Tape Cassette Bacteria Detection System
NASA Technical Reports Server (NTRS)
1973-01-01
The design, fabrication, and testing of an automatic bacteria detection system with a zero-g capability and based on the filter-capsule approach is described. This system is intended for monitoring the sterility of regenerated water in a spacecraft. The principle of detection is based on measuring the increase in chemiluminescence produced by the action of bacterial porphyrins (i.e., catalase, cytochromes, etc.) on a luminol-hydrogen peroxide mixture. Since viable as well as nonviable organisms initiate this luminescence, viable organisms are detected by comparing the signal of an incubated water sample with an unincubated control. Higher signals for the former indicate the presence of viable organisms. System features include disposable sealed sterile capsules, each containing a filter membrane, for processing discrete water samples and a tape transport for moving these capsules through a processing sequence which involves sample concentration, nutrient addition, incubation, a 4 Molar Urea wash and reaction with luminol-hydrogen peroxide in front of a photomultiplier tube. Liquids are introduced by means of a syringe needle which pierces a rubber septum contained in the wall of the capsule. Detection thresholds obtained with this unit towards E. coli and S. marcescens assuming a 400 ml water sample are indicated.
On the hitchhiker Robot Operated Materials Processing System: Experiment data system
NASA Technical Reports Server (NTRS)
Kizhner, Semion; Jenstrom, Del
1995-01-01
The Space Shuttle Discovery STS-64 mission carried the first American autonomous robot into space, the Robot Operated Materials Processing System (ROMPS). On this mission ROMPS was the only Hitchhiker experiment and had a unique opportunity to utilize all Hitchhiker space carrier capabilities. ROMPS conducted rapid thermal processing of the one hundred semiconductor material samples to study how micro gravity affects the resulting material properties. The experiment was designed, built and operated by a small GSFC team in cooperation with industry and university based principal investigators who provided the material samples and data interpretation. ROMPS' success presents some valuable lessons in such cooperation, as well as in the utilization of the Hitchhiker carrier for complex applications. The motivation of this paper is to share these lessons with the scientific community interested in attached payload experiments. ROMPS has a versatile and intelligent material processing control data system. This paper uses the ROMPS data system as the guiding thread to present the ROMPS mission experience. It presents an overview of the ROMPS experiment followed by considerations of the flight and ground data subsystems and their architecture, data products generation during mission operations, and post mission data utilization. It then presents the lessons learned from the development and operation of the ROMPS data system as well as those learned during post-flight data processing.
Morrison, Michael; Moraia, Linda Briceño; Steele, Jane C
2016-01-01
This paper describes a traceability system developed for the Stem cells for Biological Assays of Novel drugs and prediCtive toxiCology consortium. The system combines records and labels that to biological material across geographical locations and scientific processes from sample donation to induced pluripotent stem cell line. The labeling system uses a unique identification number to link every aliquot of sample at every stage of the reprogramming pathway back to the original donor. Only staff at the clinical recruitment site can reconnect the unique identification number to the identifying details of a specific donor. This ensures the system meets ethical and legal requirements for protecting privacy while allowing full traceability of biological material. The system can be adapted to other projects and for use with different primary sample types.
Optimal regulation in systems with stochastic time sampling
NASA Technical Reports Server (NTRS)
Montgomery, R. C.; Lee, P. S.
1980-01-01
An optimal control theory that accounts for stochastic variable time sampling in a distributed microprocessor based flight control system is presented. The theory is developed by using a linear process model for the airplane dynamics and the information distribution process is modeled as a variable time increment process where, at the time that information is supplied to the control effectors, the control effectors know the time of the next information update only in a stochastic sense. An optimal control problem is formulated and solved for the control law that minimizes the expected value of a quadratic cost function. The optimal cost obtained with a variable time increment Markov information update process where the control effectors know only the past information update intervals and the Markov transition mechanism is almost identical to that obtained with a known and uniform information update interval.
Amendola, Alessandra; Coen, Sabrina; Belladonna, Stefano; Pulvirenti, F Renato; Clemens, John M; Capobianchi, M Rosaria
2011-08-01
Diagnostic laboratories need automation that facilitates efficient processing and workflow management to meet today's challenges for expanding services and reducing cost, yet maintaining the highest levels of quality. Processing efficiency of two commercially available automated systems for quantifying HIV-1 and HCV RNA, Abbott m2000 system and Roche COBAS Ampliprep/COBAS TaqMan 96 (docked) systems (CAP/CTM), was evaluated in a mid/high throughput workflow laboratory using a representative daily workload of 24 HCV and 72 HIV samples. Three test scenarios were evaluated: A) one run with four batches on the CAP/CTM system, B) two runs on the Abbott m2000 and C) one run using the Abbott m2000 maxCycle feature (maxCycle) for co-processing these assays. Cycle times for processing, throughput and hands-on time were evaluated. Overall processing cycle time was 10.3, 9.1 and 7.6 h for Scenarios A), B) and C), respectively. Total hands-on time for each scenario was, in order, 100.0 (A), 90.3 (B) and 61.4 min (C). The interface of an automated analyzer to the laboratory workflow, notably system set up for samples and reagents and clean up functions, are as important as the automation capability of the analyzer for the overall impact to processing efficiency and operator hands-on time.
WAMS measurements pre-processing for detecting low-frequency oscillations in power systems
NASA Astrophysics Data System (ADS)
Kovalenko, P. Y.
2017-07-01
Processing the data received from measurement systems implies the situation when one or more registered values stand apart from the sample collection. These values are referred to as “outliers”. The processing results may be influenced significantly by the presence of those in the data sample under consideration. In order to ensure the accuracy of low-frequency oscillations detection in power systems the corresponding algorithm has been developed for the outliers detection and elimination. The algorithm is based on the concept of the irregular component of measurement signal. This component comprises measurement errors and is assumed to be Gauss-distributed random. The median filtering is employed to detect the values lying outside the range of the normally distributed measurement error on the basis of a 3σ criterion. The algorithm has been validated involving simulated signals and WAMS data as well.
The NOSAMS sample preparation laboratory in the next millenium: Progress after the WOCE program
NASA Astrophysics Data System (ADS)
Gagnon, Alan R.; McNichol, Ann P.; Donoghue, Joanne C.; Stuart, Dana R.; von Reden, Karl; Nosams
2000-10-01
Since 1991, the primary charge of the National Ocean Sciences AMS (NOSAMS) facility at the Woods Hole Oceanographic Institution has been to supply high throughput, high precision AMS 14C analyses for seawater samples collected as part of the World Ocean Circulation Experiment (WOCE). Approximately 13,000 samples taken as part of WOCE should be fully analyzed by the end of Y2K. Additional sample sources and techniques must be identified and incorporated if NOSAMS is to continue in its present operation mode. A trend in AMS today is the ability to routinely process and analyze radiocarbon samples that contain tiny amounts (<100 μg) of carbon. The capability to mass-produce small samples for 14C analysis has been recognized as a major facility goal. The installation of a new 134-position MC-SNICS ion source, which utilizes a smaller graphite target cartridge than presently used, is one step towards realizing this goal. New preparation systems constructed in the sample preparation laboratory (SPL) include an automated bank of 10 small-volume graphite reactors, an automated system to process organic carbon samples, and a multi-dimensional preparative capillary gas chromatograph (PCGC).
Sparse sampling and reconstruction for electron and scanning probe microscope imaging
Anderson, Hyrum; Helms, Jovana; Wheeler, Jason W.; Larson, Kurt W.; Rohrer, Brandon R.
2015-07-28
Systems and methods for conducting electron or scanning probe microscopy are provided herein. In a general embodiment, the systems and methods for conducting electron or scanning probe microscopy with an undersampled data set include: driving an electron beam or probe to scan across a sample and visit a subset of pixel locations of the sample that are randomly or pseudo-randomly designated; determining actual pixel locations on the sample that are visited by the electron beam or probe; and processing data collected by detectors from the visits of the electron beam or probe at the actual pixel locations and recovering a reconstructed image of the sample.
Filtration of water-sediment samples for the determination of organic compounds
Sandstrom, Mark W.
1995-01-01
This report describes the equipment and procedures used for on-site filtration of surface-water and ground-water samples for determination of organic compounds. Glass-fiber filters and a positive displacement pumping system are suitable for processing most samples for organic analyses. An optional system that uses disposable in-line membrane filters is suitable for a specific gas chromatography/mass spectrometry, selected-ion monitoring analytical method for determination of organonitrogen herbicides. General procedures to minimize contamination of the samples include preparing a clean workspace at the site, selecting appropriate sample-collection materials, and cleaning of the equipment with detergent, tap water, and methanol.
NASA Astrophysics Data System (ADS)
Mukherjee, S.; von der Heydt, M.; Hanson, C.; Jandura, L.
2009-12-01
The Mars Science Laboratory mission is scheduled to launch in 2011 with an extensive suite of in situ science instruments. Acquiring, processing and delivering appropriate samples of rock and martian regolith to the instruments is a critical component in realizing the science capability of these payload elements. However, there are a number of challenges in validating the design of these systems. In particular, differences in the environment (atmospheric pressure and composition, temperature, gravity), target materials (variation in rock and soil properties), and state of the hardware (electrical potential, particulate coatings) may effect sampling performance. To better understand the end-to-end system and allow development of mitigation strategies if necessary, early testing of high-fidelity engineering models of the hardware in the solid sample chain is being conducted. The components of the sample acquisition, processing & delivery chain that will be tested are the drill, scoop, sieves, portioners, and instrument inlet funnels. An evaluation of the environmental parameter space was conducted to identify a subset that may have significant effects on sampling performance and cannot be well bounded by analysis. Accordingly, support equipment to enable testing at Mars surface pressures (5-10 Torr), with carbon dioxide was designed and built. A description of the testing set-up, investigations, and preliminary results will be presented.
System for precise position registration
Sundelin, Ronald M.; Wang, Tong
2005-11-22
An apparatus for enabling accurate retaining of a precise position, such as for reacquisition of a microscopic spot or feature having a size of 0.1 mm or less, on broad-area surfaces after non-in situ processing. The apparatus includes a sample and sample holder. The sample holder includes a base and three support posts. Two of the support posts interact with a cylindrical hole and a U-groove in the sample to establish location of one point on the sample and a line through the sample. Simultaneous contact of the third support post with the surface of the sample defines a plane through the sample. All points of the sample are therefore uniquely defined by the sample and sample holder. The position registration system of the current invention provides accuracy, as measured in x, y repeatability, of at least 140 .mu.m.
Hyperspectral stimulated emission depletion microscopy and methods of use thereof
Timlin, Jerilyn A; Aaron, Jesse S
2014-04-01
A hyperspectral stimulated emission depletion ("STED") microscope system for high-resolution imaging of samples labeled with multiple fluorophores (e.g., two to ten fluorophores). The hyperspectral STED microscope includes a light source, optical systems configured for generating an excitation light beam and a depletion light beam, optical systems configured for focusing the excitation and depletion light beams on a sample, and systems for collecting and processing data generated by interaction of the excitation and depletion light beams with the sample. Hyperspectral STED data may be analyzed using multivariate curve resolution analysis techniques to deconvolute emission from the multiple fluorophores. The hyperspectral STED microscope described herein can be used for multi-color, subdiffraction imaging of samples (e.g., materials and biological materials) and for analyzing a tissue by Forster Resonance Energy Transfer ("FRET").
Interactive boundary delineation of agricultural lands using graphics workstations
NASA Technical Reports Server (NTRS)
Cheng, Thomas D.; Angelici, Gary L.; Slye, Robert E.; Ma, Matt
1992-01-01
A review is presented of the computer-assisted stratification and sampling (CASS) system developed to delineate the boundaries of sample units for survey procedures. CASS stratifies the sampling units by land-cover and land-use type, employing image-processing software and hardware. This procedure generates coverage areas and the boundaries of stratified sampling units that are utilized for subsequent sampling procedures from which agricultural statistics are developed.
RMP Guidance for Warehouses - Chapter 5: Management System
Your management system should oversee the implementation of the risk management program elements, and designate and assign responsibility in order to make process safety a constant priority. Includes sample documentation.
Identification and quantification of pathogenic helminth eggs using a digital image system.
Jiménez, B; Maya, C; Velásquez, G; Torner, F; Arambula, F; Barrios, J A; Velasco, M
2016-07-01
A system was developed to identify and quantify up to seven species of helminth eggs (Ascaris lumbricoides -fertile and unfertile eggs-, Trichuris trichiura, Toxocara canis, Taenia saginata, Hymenolepis nana, Hymenolepis diminuta, and Schistosoma mansoni) in wastewater using different image processing tools and pattern recognition algorithms. The system was developed in three stages. Version one was used to explore the viability of the concept of identifying helminth eggs through an image processing system, while versions 2 and 3 were used to improve its efficiency. The system development was based on the analysis of different properties of helminth eggs in order to discriminate them from other objects in samples processed using the conventional United States Environmental Protection Agency (US EPA) technique to quantify helminth eggs. The system was tested, in its three stages, considering two parameters: specificity (capacity to discriminate between species of helminth eggs and other objects) and sensitivity (capacity to correctly classify and identify the different species of helminth eggs). The final version showed a specificity of 99% while the sensitivity varied between 80 and 90%, depending on the total suspended solids content of the wastewater samples. To achieve such values in samples with total suspended solids (TSS) above 150 mg/L, it is recommended to dilute the concentrated sediment just before taking the images under the microscope. The system allows the helminth eggs most commonly found in wastewater to be reliably and uniformly detected and quantified. In addition, it provides the total number of eggs as well as the individual number by species, and for Ascaris lumbricoides it differentiates whether or not the egg is fertile. The system only requires basically trained technicians to prepare the samples, as for visual identification there is no need for highly trained personnel. The time required to analyze each image is less than a minute. This system could be used in central analytical laboratories providing a remote analysis service. Copyright © 2016 The Authors. Published by Elsevier Inc. All rights reserved.
Shubhakar, Archana; Kalla, Rahul; Nimmo, Elaine R.; Fernandes, Daryl L.; Satsangi, Jack; Spencer, Daniel I. R.
2015-01-01
Introduction Serum N-glycans have been identified as putative biomarkers for numerous diseases. The impact of different serum sample tubes and processing methods on N-glycan analysis has received relatively little attention. This study aimed to determine the effect of different sample tubes and processing methods on the whole serum N-glycan profile in both health and disease. A secondary objective was to describe a robot automated N-glycan release, labeling and cleanup process for use in a biomarker discovery system. Methods 25 patients with active and quiescent inflammatory bowel disease and controls had three different serum sample tubes taken at the same draw. Two different processing methods were used for three types of tube (with and without gel-separation medium). Samples were randomised and processed in a blinded fashion. Whole serum N-glycan release, 2-aminobenzamide labeling and cleanup was automated using a Hamilton Microlab STARlet Liquid Handling robot. Samples were analysed using a hydrophilic interaction liquid chromatography/ethylene bridged hybrid(BEH) column on an ultra-high performance liquid chromatography instrument. Data were analysed quantitatively by pairwise correlation and hierarchical clustering using the area under each chromatogram peak. Qualitatively, a blinded assessor attempted to match chromatograms to each individual. Results There was small intra-individual variation in serum N-glycan profiles from samples collected using different sample processing methods. Intra-individual correlation coefficients were between 0.99 and 1. Unsupervised hierarchical clustering and principal coordinate analyses accurately matched samples from the same individual. Qualitative analysis demonstrated good chromatogram overlay and a blinded assessor was able to accurately match individuals based on chromatogram profile, regardless of disease status. Conclusions The three different serum sample tubes processed using the described methods cause minimal inter-individual variation in serum whole N-glycan profile when processed using an automated workstream. This has important implications for N-glycan biomarker discovery studies using different serum processing standard operating procedures. PMID:25831126
Ventham, Nicholas T; Gardner, Richard A; Kennedy, Nicholas A; Shubhakar, Archana; Kalla, Rahul; Nimmo, Elaine R; Fernandes, Daryl L; Satsangi, Jack; Spencer, Daniel I R
2015-01-01
Serum N-glycans have been identified as putative biomarkers for numerous diseases. The impact of different serum sample tubes and processing methods on N-glycan analysis has received relatively little attention. This study aimed to determine the effect of different sample tubes and processing methods on the whole serum N-glycan profile in both health and disease. A secondary objective was to describe a robot automated N-glycan release, labeling and cleanup process for use in a biomarker discovery system. 25 patients with active and quiescent inflammatory bowel disease and controls had three different serum sample tubes taken at the same draw. Two different processing methods were used for three types of tube (with and without gel-separation medium). Samples were randomised and processed in a blinded fashion. Whole serum N-glycan release, 2-aminobenzamide labeling and cleanup was automated using a Hamilton Microlab STARlet Liquid Handling robot. Samples were analysed using a hydrophilic interaction liquid chromatography/ethylene bridged hybrid(BEH) column on an ultra-high performance liquid chromatography instrument. Data were analysed quantitatively by pairwise correlation and hierarchical clustering using the area under each chromatogram peak. Qualitatively, a blinded assessor attempted to match chromatograms to each individual. There was small intra-individual variation in serum N-glycan profiles from samples collected using different sample processing methods. Intra-individual correlation coefficients were between 0.99 and 1. Unsupervised hierarchical clustering and principal coordinate analyses accurately matched samples from the same individual. Qualitative analysis demonstrated good chromatogram overlay and a blinded assessor was able to accurately match individuals based on chromatogram profile, regardless of disease status. The three different serum sample tubes processed using the described methods cause minimal inter-individual variation in serum whole N-glycan profile when processed using an automated workstream. This has important implications for N-glycan biomarker discovery studies using different serum processing standard operating procedures.
Filling in the Gaps: Xenoliths in Meteorites are Samples of "Missing" Asteroid Lithologies
NASA Technical Reports Server (NTRS)
Zolensky, Mike
2016-01-01
We know that the stones that fall to earth as meteorites are not representative of the full diversity of small solar system bodies, because of the peculiarities of the dynamical processes that send material into Earth-crossing paths [1] which result in severe selection biases. Thus, the bulk of the meteorites that fall are insufficient to understand the full range of early solar system processes. However, the situation is different for pebble- and smaller-sized objects that stream past the giant planets and asteroid belts into the inner solar system in a representative manner. Thus, micrometeorites and interplanetary dust particles have been exploited to permit study of objects that do not provide meteorites to earth. However, there is another population of materials that sample a larger range of small solar system bodies, but which have received little attention - pebble-sized foreign clasts in meteorites (also called xenoliths, dark inclusions, clasts, etc.). Unfortunately, most previous studies of these clasts have been misleading, in that these objects have simply been identified as pieces of CM or CI chondrites. In our work we have found this to be generally erroneous, and that CM and especially CI clasts are actually rather rare. We therefore test the hypothesis that these clasts sample the full range of small solar system bodies. We have located and obtained samples of clasts in 81 different meteorites, and have begun a thorough characterization of the bulk compositions, mineralogies, petrographies, and organic compositions of this unique sample set. In addition to the standard e-beam analyses, recent advances in technology now permit us to measure bulk O isotopic compositions, and major- though trace-element compositions of the sub-mm-sized discrete clasts. Detailed characterization of these clasts permit us to explore the full range of mineralogical and petrologic processes in the early solar system, including the nature of fluids in the Kuiper belt and the outer main asteroid belt, as revealed by the mineralogy of secondary phases.
Hyperspectral Imaging and Obstacle Detection for Robotics Navigation
2005-09-01
anatomy and diffraction process. 17 3.3 Technical Specifications of the System A. Brimrose AOTF Video Adaptor Specifications: Material TeO2 Active...sampled from glass case on person 2’s belt 530 pixels 20 pick-up white sampled from body panels of pick-up 600 pixels 21 pick-up blue sampled from
Weber, J. K. R.; Tamalonis, A.; Benmore, C. J.; ...
2016-07-01
We integrated an aerodynamic levitator with carbon dioxide laser beam heating with a hermetically sealed controlled atmosphere chamber and sample handling mechanism. The system enabled containment of radioactive samples and control of the process atmosphere chemistry. Furthermore, the chamber was typically operated at a pressure of approximately 0.9 bars to ensure containment of the materials being processed. Samples 2.5-3 mm in diameter were levitated in flowing gas to achieve containerless conditions. Levitated samples were heated to temperatures of up to 3500 °C with a partially focused carbon dioxide laser beam. Sample temperature was measured using an optical pyrometer. The samplemore » environment was integrated with a high energy (100 keV) x-ray synchrotron beamline to enable in situ structure measurements to be made on levitated samples as they were heated, melted, and supercooled. Our system was controlled from outside the x-ray beamline hutch by using a LabVIEW program. Measurements have been made on hot solid and molten uranium dioxide and binary uranium dioxide-zirconium dioxide compositions.« less
NASA Astrophysics Data System (ADS)
Hayasaki, Yoshio
2017-02-01
Femtosecond laser processing is a promising tool for fabricating novel and useful structures on the surfaces of and inside materials. An enormous number of pulse irradiation points will be required for fabricating actual structures with millimeter scale, and therefore, the throughput of femtosecond laser processing must be improved for practical adoption of this technique. One promising method to improve throughput is parallel pulse generation based on a computer-generated hologram (CGH) displayed on a spatial light modulator (SLM), a technique called holographic femtosecond laser processing. The holographic method has the advantages such as high throughput, high light use efficiency, and variable, instantaneous, and 3D patterning. Furthermore, the use of an SLM gives an ability to correct unknown imperfections of the optical system and inhomogeneity in a sample using in-system optimization of the CGH. Furthermore, the CGH can adaptively compensate in response to dynamic unpredictable mechanical movements, air and liquid disturbances, a shape variation and deformation of the target sample, as well as adaptive wavefront control for environmental changes. Therefore, it is a powerful tool for the fabrication of biological cells and tissues, because they have free form, variable, and deformable structures. In this paper, we present the principle and the experimental setup of holographic femtosecond laser processing, and the effective way for processing the biological sample. We demonstrate the femtosecond laser processing of biological materials and the processing properties.
Almeida, Jamie L.; Wang, Lili; Morrow, Jayne B.; Cole, Kenneth D.
2006-01-01
Bacillus anthracis spores have been used as biological weapons and the possibility of their further use requires surveillance systems that can accurately and reliably detect their presence in the environment. These systems must collect samples from a variety of matrices, process the samples, and detect the spores. The processing of the sample may include removal of inhibitors, concentration of the target, and extraction of the target in a form suitable for detection. Suitable reference materials will allow the testing of each of these steps to determine the sensitivity and specificity of the detection systems. The development of uniform and well-characterized reference materials will allow the comparison of different devices and technologies as well as assure the continued performance of detection systems. This paper discusses the special requirements of reference materials for Bacillus anthracis spores that could be used for testing detection systems. The detection of Bacillus anthracis spores is based on recognition of specific characteristics (markers) on either the spore surface or in the nucleic acids (DNA). We have reviewed the specific markers and their relevance to characterization of reference materials. We have also included the approach for the characterization of candidate reference materials that we are developing at the NIST laboratories. Additional applications of spore reference materials would include testing sporicidal treatments, techniques for sampling the environment, and remediation of spore-contaminated environments. PMID:27274929
High Resolution X-Ray Micro-CT of Ultra-Thin Wall Space Components
NASA Technical Reports Server (NTRS)
Roth, Don J.; Rauser, R. W.; Bowman, Randy R.; Bonacuse, Peter; Martin, Richard E.; Locci, I. E.; Kelley, M.
2012-01-01
A high resolution micro-CT system has been assembled and is being used to provide optimal characterization for ultra-thin wall space components. The Glenn Research Center NDE Sciences Team, using this CT system, has assumed the role of inspection vendor for the Advanced Stirling Convertor (ASC) project at NASA. This article will discuss many aspects of the development of the CT scanning for this type of component, including CT system overview; inspection requirements; process development, software utilized and developed to visualize, process, and analyze results; calibration sample development; results on actual samples; correlation with optical/SEM characterization; CT modeling; and development of automatic flaw recognition software. Keywords: Nondestructive Evaluation, NDE, Computed Tomography, Imaging, X-ray, Metallic Components, Thin Wall Inspection
NASA Technical Reports Server (NTRS)
Meisch, A. J.
1972-01-01
Data for the system n-pentane/n-heptane on porous Chromosorb-102 adsorbent were obtained at 150, 175, and 200 C for mixtures containing zero to 100% n-pentane by weight. Prior results showing limitations on superposition of pure component data to predict multicomponent chromatograms were verified. The thermodynamic parameter MR0 was found to be a linear function of sample composition. A nonporous adsorbent failed to separate the system because of large input sample dispersions. A proposed automated data processing scheme involving magnetic tape recording of the detector signals and processing by a minicomputer was rejected because of resolution limitations of the available a/d converters. Preliminary data on porosity and pore size distributions of the adsorbents were obtained.
Markert, Sven; Joeris, Klaus
2017-01-01
We developed an automated microtiter plate (MTP)-based system for suspension cell culture to meet the increased demands for miniaturized high throughput applications in biopharmaceutical process development. The generic system is based on off-the-shelf commercial laboratory automation equipment and is able to utilize MTPs of different configurations (6-24 wells per plate) in orbital shaken mode. The shaking conditions were optimized by Computational Fluid Dynamics simulations. The fully automated system handles plate transport, seeding and feeding of cells, daily sampling, and preparation of analytical assays. The integration of all required analytical instrumentation into the system enables a hands-off operation which prevents bottlenecks in sample processing. The modular set-up makes the system flexible and adaptable for a continuous extension of analytical parameters and add-on components. The system proved suitable as screening tool for process development by verifying the comparability of results for the MTP-based system and bioreactors regarding profiles of viable cell density, lactate, and product concentration of CHO cell lines. These studies confirmed that 6 well MTPs as well as 24 deepwell MTPs were predictive for a scale up to a 1000 L stirred tank reactor (scale factor 1:200,000). Applying the established cell culture system for automated media blend screening in late stage development, a 22% increase in product yield was achieved in comparison to the reference process. The predicted product increase was subsequently confirmed in 2 L bioreactors. Thus, we demonstrated the feasibility of the automated MTP-based cell culture system for enhanced screening and optimization applications in process development and identified further application areas such as process robustness. The system offers a great potential to accelerate time-to-market for new biopharmaceuticals. Biotechnol. Bioeng. 2017;114: 113-121. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
A Versatile High-Vacuum Cryo-transfer System for Cryo-microscopy and Analytics
Tacke, Sebastian; Krzyzanek, Vladislav; Nüsse, Harald; Wepf, Roger Albert; Klingauf, Jürgen; Reichelt, Rudolf
2016-01-01
Cryogenic microscopy methods have gained increasing popularity, as they offer an unaltered view on the architecture of biological specimens. As a prerequisite, samples must be handled under cryogenic conditions below their recrystallization temperature, and contamination during sample transfer and handling must be prevented. We present a high-vacuum cryo-transfer system that streamlines the entire handling of frozen-hydrated samples from the vitrification process to low temperature imaging for scanning transmission electron microscopy and transmission electron microscopy. A template for cryo-electron microscopy and multimodal cryo-imaging approaches with numerous sample transfer steps is presented. PMID:26910419
DOE Office of Scientific and Technical Information (OSTI.GOV)
Riot, Vincent J.
The present disclosure provides a system and a method for measuring fluorescence of a sample. The sample may be a polymerase-chain-reaction (PCR) array, a loop-mediated-isothermal amplification array, etc. LEDs are used to excite the sample, and a photodiode is used to collect the sample's fluorescence. An electronic offset signal is used to reduce the effects of background fluorescence and the noises from the measurement system. An integrator integrates the difference between the output of the photodiode and the electronic offset signal over a given period of time. The resulting integral is then converted into digital domain for further processing andmore » storage.« less
Garty, Guy; Chen, Youhua; Turner, Helen C; Zhang, Jian; Lyulko, Oleksandra V; Bertucci, Antonella; Xu, Yanping; Wang, Hongliang; Simaan, Nabil; Randers-Pehrson, Gerhard; Lawrence Yao, Y; Brenner, David J
2011-08-01
Over the past five years the Center for Minimally Invasive Radiation Biodosimetry at Columbia University has developed the Rapid Automated Biodosimetry Tool (RABiT), a completely automated, ultra-high throughput biodosimetry workstation. This paper describes recent upgrades and reliability testing of the RABiT. The RABiT analyses fingerstick-derived blood samples to estimate past radiation exposure or to identify individuals exposed above or below a cut-off dose. Through automated robotics, lymphocytes are extracted from fingerstick blood samples into filter-bottomed multi-well plates. Depending on the time since exposure, the RABiT scores either micronuclei or phosphorylation of the histone H2AX, in an automated robotic system, using filter-bottomed multi-well plates. Following lymphocyte culturing, fixation and staining, the filter bottoms are removed from the multi-well plates and sealed prior to automated high-speed imaging. Image analysis is performed online using dedicated image processing hardware. Both the sealed filters and the images are archived. We have developed a new robotic system for lymphocyte processing, making use of an upgraded laser power and parallel processing of four capillaries at once. This system has allowed acceleration of lymphocyte isolation, the main bottleneck of the RABiT operation, from 12 to 2 sec/sample. Reliability tests have been performed on all robotic subsystems. Parallel handling of multiple samples through the use of dedicated, purpose-built, robotics and high speed imaging allows analysis of up to 30,000 samples per day.
The Informatics Challenges Facing Biobanks: A Perspective from a United Kingdom Biobanking Network
Groves, Martin; Jordan, Lee B.; Stobart, Hilary; Purdie, Colin A.; Thompson, Alastair M
2015-01-01
The challenges facing biobanks are changing from simple collections of materials to quality-assured fit-for-purpose clinically annotated samples. As a result, informatics awareness and capabilities of a biobank are now intrinsically related to quality. A biobank may be considered a data repository, in the form of raw data (the unprocessed samples), data surrounding the samples (processing and storage conditions), supplementary data (such as clinical annotations), and an increasing ethical requirement for biobanks to have a mechanism for researchers to return their data. The informatics capabilities of a biobank are no longer simply knowing sample locations; instead the capabilities will become a distinguishing factor in the ability of a biobank to provide appropriate samples. There is an increasing requirement for biobanking systems (whether in-house or commercially sourced) to ensure the informatics systems stay apace with the changes being experienced by the biobanking community. In turn, there is a requirement for the biobanks to have a clear informatics policy and directive that is embedded into the wider decision making process. As an example, the Breast Cancer Campaign Tissue Bank in the UK was a collaboration between four individual and diverse biobanks in the UK, and an informatics platform has been developed to address the challenges of running a distributed network. From developing such a system there are key observations about what can or cannot be achieved by informatics in isolation. This article will highlight some of the lessons learned during this development process. PMID:26418270
Garty, Guy; Chen, Youhua; Turner, Helen; Zhang, Jian; Lyulko, Oleksandra; Bertucci, Antonella; Xu, Yanping; Wang, Hongliang; Simaan, Nabil; Randers-Pehrson, Gerhard; Yao, Y. Lawrence; Brenner, David J.
2011-01-01
Purpose Over the past five years the Center for Minimally Invasive Radiation Biodosimetry at Columbia University has developed the Rapid Automated Biodosimetry Tool (RABiT), a completely automated, ultra-high throughput biodosimetry workstation. This paper describes recent upgrades and reliability testing of the RABiT. Materials and methods The RABiT analyzes fingerstick-derived blood samples to estimate past radiation exposure or to identify individuals exposed above or below a cutoff dose. Through automated robotics, lymphocytes are extracted from fingerstick blood samples into filter-bottomed multi-well plates. Depending on the time since exposure, the RABiT scores either micronuclei or phosphorylation of the histone H2AX, in an automated robotic system, using filter-bottomed multi-well plates. Following lymphocyte culturing, fixation and staining, the filter bottoms are removed from the multi-well plates and sealed prior to automated high-speed imaging. Image analysis is performed online using dedicated image processing hardware. Both the sealed filters and the images are archived. Results We have developed a new robotic system for lymphocyte processing, making use of an upgraded laser power and parallel processing of four capillaries at once. This system has allowed acceleration of lymphocyte isolation, the main bottleneck of the RABiT operation, from 12 to 2 sec/sample. Reliability tests have been performed on all robotic subsystems. Conclusions Parallel handling of multiple samples through the use of dedicated, purpose-built, robotics and high speed imaging allows analysis of up to 30,000 samples per day. PMID:21557703
A wireless data acquisition system for acoustic emission testing
NASA Astrophysics Data System (ADS)
Zimmerman, A. T.; Lynch, J. P.
2013-01-01
As structural health monitoring (SHM) systems have seen increased demand due to lower costs and greater capabilities, wireless technologies have emerged that enable the dense distribution of transducers and the distributed processing of sensor data. In parallel, ultrasonic techniques such as acoustic emission (AE) testing have become increasingly popular in the non-destructive evaluation of materials and structures. These techniques, which involve the analysis of frequency content between 1 kHz and 1 MHz, have proven effective in detecting the onset of cracking and other early-stage failure in active structures such as airplanes in flight. However, these techniques typically involve the use of expensive and bulky monitoring equipment capable of accurately sensing AE signals at sampling rates greater than 1 million samples per second. In this paper, a wireless data acquisition system is presented that is capable of collecting, storing, and processing AE data at rates of up to 20 MHz. Processed results can then be wirelessly transmitted in real-time, creating a system that enables the use of ultrasonic techniques in large-scale SHM systems.
Han, Liang-Feng; Plummer, Niel; Aggarwal, Pradeep
2012-01-01
A graphical method is described for identifying geochemical reactions needed in the interpretation of radiocarbon age in groundwater systems. Graphs are constructed by plotting the measured 14C, δ13C, and concentration of dissolved inorganic carbon and are interpreted according to specific criteria to recognize water samples that are consistent with a wide range of processes, including geochemical reactions, carbon isotopic exchange, 14C decay, and mixing of waters. The graphs are used to provide a qualitative estimate of radiocarbon age, to deduce the hydrochemical complexity of a groundwater system, and to compare samples from different groundwater systems. Graphs of chemical and isotopic data from a series of previously-published groundwater studies are used to demonstrate the utility of the approach. Ultimately, the information derived from the graphs is used to improve geochemical models for adjustment of radiocarbon ages in groundwater systems.
Greenspoon, Susan A; Ban, Jeffrey D; Sykes, Karen; Ballard, Elizabeth J; Edler, Shelley S; Baisden, Melissa; Covington, Brian L
2004-01-01
Robotic systems are commonly utilized for the extraction of database samples. However, the application of robotic extraction to forensic casework samples is a more daunting task. Such a system must be versatile enough to accommodate a wide range of samples that may contain greatly varying amounts of DNA, but it must also pose no more risk of contamination than the manual DNA extraction methods. This study demonstrates that the BioMek 2000 Laboratory Automation Workstation, used in combination with the DNA IQ System, is versatile enough to accommodate the wide range of samples typically encountered by a crime laboratory. The use of a silica coated paramagnetic resin, as with the DNA IQ System, facilitates the adaptation of an open well, hands off, robotic system to the extraction of casework samples since no filtration or centrifugation steps are needed. Moreover, the DNA remains tightly coupled to the silica coated paramagnetic resin for the entire process until the elution step. A short pre-extraction incubation step is necessary prior to loading samples onto the robot and it is at this step that most modifications are made to accommodate the different sample types and substrates commonly encountered with forensic evidentiary samples. Sexual assault (mixed stain) samples, cigarette butts, blood stains, buccal swabs, and various tissue samples were successfully extracted with the BioMek 2000 Laboratory Automation Workstation and the DNA IQ System, with no evidence of contamination throughout the extensive validation studies reported here.
Saini, Parmesh K; Marks, Harry M; Dreyfuss, Moshe S; Evans, Peter; Cook, L Victor; Dessai, Uday
2011-08-01
Measuring commonly occurring, nonpathogenic organisms on poultry products may be used for designing statistical process control systems that could result in reductions of pathogen levels. The extent of pathogen level reduction that could be obtained from actions resulting from monitoring these measurements over time depends upon the degree of understanding cause-effect relationships between processing variables, selected output variables, and pathogens. For such measurements to be effective for controlling or improving processing to some capability level within the statistical process control context, sufficiently frequent measurements would be needed to help identify processing deficiencies. Ultimately the correct balance of sampling and resources is determined by those characteristics of deficient processing that are important to identify. We recommend strategies that emphasize flexibility, depending upon sampling objectives. Coupling the measurement of levels of indicator organisms with practical emerging technologies and suitable on-site platforms that decrease the time between sample collections and interpreting results would enhance monitoring process control.
NASA Technical Reports Server (NTRS)
Alfven, H.; Arrhenius, G.
1976-01-01
The origin and evolution of the solar system are analyzed. Physical processes are first discussed, followed by experimental studies of plasma-solid reactions and chemical and mineralogical analyses of meteorites and lunar and terrestrial samples.
Elbeik, Tarek; Loftus, Richard A; Beringer, Scott
2007-11-01
Labor, supply and waste were evaluated for HIV-1 and HCV bDNA on the semi-automated System 340 bDNA Analyzer and the automated VERSANT 440 Molecular System (V440). HIV-1 sample processing was evaluated using a 24- and 48-position centrifuge rotor. Vigilance time (hands-on manipulations plus incubation time except initial target hybridization) and disposables were approximately 37 and 12% lower for HIV-1, and 64 and 31% lower for HCV bDNA, respectively, with V440. Biohazardous solid waste was approximately twofold lower for both assays and other waste types were the same for either assay on both platforms. HIV-1 sample processing vigilance time for the 48-position rotor was reduced by 2 h. V440 provides cost savings and improved workflow.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Demos, Stavros; Levenson, Richard
The present disclosure relates to a method for analyzing tissue specimens. In one implementation the method involves obtaining a tissue sample and exposing the sample to one or more fluorophores as contrast agents to enhance contrast of subcellular compartments of the tissue sample. The tissue sample is illuminated by an ultraviolet (UV) light having a wavelength between about 200 nm to about 400 nm, with the wavelength being selected to result in penetration to only a specified depth below a surface of the tissue sample. Inter-image operations between images acquired under different imaging parameters allow for improvement of the imagemore » quality via removal of unwanted image components. A microscope may be used to image the tissue sample and provide the image to an image acquisition system that makes use of a camera. The image acquisition system may create a corresponding image that is transmitted to a display system for processing and display.« less
Influence of Landscape Morphology and Vegetation Cover on the Sampling of Mixed Igneous Bodies
NASA Astrophysics Data System (ADS)
Perugini, Diego; Petrelli, Maurizio; Poli, Giampiero
2010-05-01
A plethora of evidence indicates that magma mixing processes can take place at any evolutionary stage of magmatic systems and that they are extremely common in both plutonic and volcanic environments (e.g. Bateman, 1995). Furthermore, recent studies have shown that the magma mixing process is governed by chaotic dynamics whose evolution in space and time generates complex compositional patterns that can span several length scales producing fractal domains (e.g. Perugini et al., 2003). The fact that magma mixing processes can produce igneous bodies exhibiting a large compositional complexity brings up the key question about the potential pitfalls that may be associated with the sampling of these systems for petrological studies. In particular, since commonly only exiguous portions of the whole magmatic system are available as outcrops for sampling, it is important to address the point whether the sampling may be considered representative of the complexity of the magmatic system. We attempt to address this crucial point by performing numerical simulations of chaotic magma mixing processes in 3D. The numerical system used in the simulations is the so-called ABC (Arnold-Beltrami-Childress) flow (e.g. Galluccio and Vulpiani, 1994), which is able to generate the contemporaneous occurrence of chaotic and regular streamlines in which the mixing efficiency is differently modulated. This numerical system has already been successfully utilized as a kinematic template to reproduce magma mixing structures observed on natural outcrops (Perugini et al., 2007). The best conditions for sampling are evaluated considering different landscape morphologies and percentages of vegetation cover. In particular, synthetic landscapes with different degree of roughness are numerically reproduced using the Random Mid-point Displacement Method (RMDM; e.g. Fournier et al., 1982) in two dimensions and superimposed to the compositional fields generated by the magma mixing simulation. Vegetation cover is generated using a random Brownian motion process in 2D. Such an approach allows us to produce vegetation patches that closely match the general topology of natural vegetation (e.g., Mandelbrot, 1982). Results show that the goodness of sampling is strongly dependant on the roughness of the landscape, with highly irregular morphologies being the best candidates to give the most complete information on the whole magma body. Conversely, sampling on flat or nearly flat surfaces should be avoided because they may contain misleading information about the magmatic system. Contrary to common sense, vegetation cover does not appear to significantly influence the representativeness of sampling if sample collection occurs on topographically irregular outcrops. Application of the proposed method for sampling area selection is straightforward. The irregularity of natural landscapes and the percentage of vegetation can be estimated by using natural landscapes extracted from digital elevation models (DEM) of the Earth's surface and satellite images by employing a variety of methods (e.g., Develi and Babadagli, 1998), thus giving one the opportunity to select a priori the best outcrops for sampling. References Bateman R (1995) The interplay between crystallization, replenishment and hybridization in large felsic magma chambers. Earth Sci Rev 39: 91-106 Develi K, Babadagli T (1998) Quantfication of natural fracture surfaces using fractal geometry. Math Geol 30: 971-998 Fournier A, Fussel D, Carpenter L (1982) Computer rendering of stochastic models. Comm ACM 25: 371-384 Galluccio S, Vulpiani A (1994) Stretching of material lines and surfaces in systems with Lagrangian chaos. Physica A 212: 75-98 Mandelbrot BB (1982) The fractal geometry of nature. W. H. Freeman, San Francisco Perugini D, Petrelli M, Poli G (2007) A Virtual Voyage through 3D Structures Generated by Chaotic Mixing of Magmas and Numerical Simulations: a New Approach for Understanding Spatial and Temporal Complexity of Magma Dynamics, Visual Geosciences, 10.1007/s10069-006-0004-x Perugini D, Poli G, Mazzuoli R (2003) Chaotic advection, fractals and diffusion during mixing of magmas: evidences from lava flows. J Volcanol Geotherm Res 124: 255-279
40 CFR 86.160-00 - Exhaust emission test procedure for SC03 emissions.
Code of Federal Regulations, 2010 CFR
2010-07-01
... simulates testing in an environmental test cell (see § 86.162-00 (a) for a discussion of simulation... exhaust and dilution air bag samples to the analytical system and process the samples according to § 86...
An integratable microfluidic cartridge for forensic swab samples lysis.
Yang, Jianing; Brooks, Carla; Estes, Matthew D; Hurth, Cedric M; Zenhausern, Frederic
2014-01-01
Fully automated rapid forensic DNA analysis requires integrating several multistep processes onto a single microfluidic platform, including substrate lysis, extraction of DNA from the released lysate solution, multiplexed PCR amplification of STR loci, separation of PCR products by capillary electrophoresis, and analysis for allelic peak calling. Over the past several years, most of the rapid DNA analysis systems developed started with the reference swab sample lysate and involved an off-chip lysis of collected substrates. As a result of advancement in technology and chemistry, addition of a microfluidic module for swab sample lysis has been achieved in a few of the rapid DNA analysis systems. However, recent reports on integrated rapid DNA analysis systems with swab-in and answer-out capability lack any quantitative and qualitative characterization of the swab-in sample lysis module, which is important for downstream forensic sample processing. Maximal collection and subsequent recovery of the biological material from the crime scene is one of the first and critical steps in forensic DNA technology. Herein we present the design, fabrication and characterization of an integratable swab lysis cartridge module and the test results obtained from different types of commonly used forensic swab samples, including buccal, saliva, and blood swab samples, demonstrating the compatibility with different downstream DNA extraction chemistries. This swab lysis cartridge module is easy to operate, compatible with both forensic and microfluidic requirements, and ready to be integrated with our existing automated rapid forensic DNA analysis system. Following the characterization of the swab lysis module, an integrated run from buccal swab sample-in to the microchip CE electropherogram-out was demonstrated on the integrated prototype instrument. Therefore, in this study, we demonstrate that this swab lysis cartridge module is: (1) functionally, comparable with routine benchtop lysis, (2) compatible with various types of swab samples and chemistries, and (3) integratable to achieve a micro total analysis system (μTAS) for rapid DNA analysis. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.
Experience with soluble neutron poisons for criticality control at ICPP
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wilson, R.E.; Mortimer, S.R.
1978-01-01
Soluble neutron poisons assure criticality control in two of the headend fuel reprocessing systems at the Idaho Chemical Processing Plant. Soluble poisons have been used successfully since 1964 and will be employed in the projected new headend processes. The use of soluble poisons (1) greatly increases the process output (2) allows versatility in the size of fuel assemblies processed and (3) allows the practical reprocessing of some fuels. The safety limit for all fluids entering the U-Zr alloy dissolver is 3.6 g/liter boron. To allow for possible deviations in the measurement systems and drift between analytical sampling periods, the standardmore » practice is to use 3.85 g/liter boron as the lower limit. This dissolver has had 4000 successful hours of operation using soluble poisons. The electrolytic dissolution process depends on soluble gadolinium for criticality safety. This system is used to process high enriched uranium clad in stainless steel. Electrolytic dissolution takes advantage of the anodic corrosion that occurs when a large electrical current is passed through the fuel elements in a corrosive environment. Three control methods are used on each headend system. First, the poison is mixed according to standard operating procedures and the measurements are affirmed by the operator's supervisor. Second, the poisoned solution is stirred, sampled, analyzed, and the analysis reported while still in the mix tank. Finally, a Nuclear Poison Detection System (NPDS) must show an acceptable poison concentration before the solution can be transferred. The major disadvantage of using soluble poisons is the need for very sophisticated control systems and procedures, which require extensive checkout. The need for a poisoned primary heating and cooling system means a secondary system is needed as well. Experience has shown, however, that production enhancement more than makes up for the problems.« less
BPSK Demodulation Using Digital Signal Processing
NASA Technical Reports Server (NTRS)
Garcia, Thomas R.
1996-01-01
A digital communications signal is a sinusoidal waveform that is modified by a binary (digital) information signal. The sinusoidal waveform is called the carrier. The carrier may be modified in amplitude, frequency, phase, or a combination of these. In this project a binary phase shift keyed (BPSK) signal is the communication signal. In a BPSK signal the phase of the carrier is set to one of two states, 180 degrees apart, by a binary (i.e., 1 or 0) information signal. A digital signal is a sampled version of a "real world" time continuous signal. The digital signal is generated by sampling the continuous signal at discrete points in time. The rate at which the signal is sampled is called the sampling rate (f(s)). The device that performs this operation is called an analog-to-digital (A/D) converter or a digitizer. The digital signal is composed of the sequence of individual values of the sampled BPSK signal. Digital signal processing (DSP) is the modification of the digital signal by mathematical operations. A device that performs this processing is called a digital signal processor. After processing, the digital signal may then be converted back to an analog signal using a digital-to-analog (D/A) converter. The goal of this project is to develop a system that will recover the digital information from a BPSK signal using DSP techniques. The project is broken down into the following steps: (1) Development of the algorithms required to demodulate the BPSK signal; (2) Simulation of the system; and (3) Implementation a BPSK receiver using digital signal processing hardware.
Automated mass spectrometer analysis system
NASA Technical Reports Server (NTRS)
Giffin, Charles E. (Inventor); Kuppermann, Aron (Inventor); Dreyer, William J. (Inventor); Boettger, Heinz G. (Inventor)
1982-01-01
An automated mass spectrometer analysis system is disclosed, in which samples are automatically processed in a sample processor and converted into volatilizable samples, or their characteristic volatilizable derivatives. Each volatilizable sample is sequentially volatilized and analyzed in a double focusing mass spectrometer, whose output is in the form of separate ion beams all of which are simultaneously focused in a focal plane. Each ion beam is indicative of a different sample component or different fragments of one or more sample components and the beam intensity is related to the relative abundance of the sample component. The system includes an electro-optical ion detector which automatically and simultaneously converts the ion beams, first into electron beams which in turn produce a related image which is transferred to the target of a vilicon unit. The latter converts the images into electrical signals which are supplied to a data processor, whose output is a list of the components of the analyzed sample and their abundances. The system is under the control of a master control unit, which in addition to monitoring and controlling various power sources, controls the automatic operation of the system under expected and some unexpected conditions and further protects various critical parts of the system from damage due to particularly abnormal conditions.
Automated mass spectrometer analysis system
NASA Technical Reports Server (NTRS)
Boettger, Heinz G. (Inventor); Giffin, Charles E. (Inventor); Dreyer, William J. (Inventor); Kuppermann, Aron (Inventor)
1978-01-01
An automated mass spectrometer analysis system is disclosed, in which samples are automatically processed in a sample processor and converted into volatilizable samples, or their characteristic volatilizable derivatives. Each volatizable sample is sequentially volatilized and analyzed in a double focusing mass spectrometer, whose output is in the form of separate ion beams all of which are simultaneously focused in a focal plane. Each ion beam is indicative of a different sample component or different fragments of one or more sample components and the beam intensity is related to the relative abundance of the sample component. The system includes an electro-optical ion detector which automatically and simultaneously converts the ion beams, first into electron beams which in turn produce a related image which is transferred to the target of a vidicon unit. The latter converts the images into electrical signals which are supplied to a data processor, whose output is a list of the components of the analyzed sample and their abundances. The system is under the control of a master control unit, which in addition to monitoring and controlling various power sources, controls the automatic operation of the system under expected and some unexpected conditions and further protects various critical parts of the system from damage due to particularly abnormal conditions.
A novel PMT test system based on waveform sampling
NASA Astrophysics Data System (ADS)
Yin, S.; Ma, L.; Ning, Z.; Qian, S.; Wang, Y.; Jiang, X.; Wang, Z.; Yu, B.; Gao, F.; Zhu, Y.; Wang, Z.
2018-01-01
Comparing with the traditional test system based on a QDC and TDC and scaler, a test system based on waveform sampling is constructed for signal sampling of the 8"R5912 and the 20"R12860 Hamamatsu PMT in different energy states from single to multiple photoelectrons. In order to achieve high throughput and to reduce the dead time in data processing, the data acquisition software based on LabVIEW is developed and runs with a parallel mechanism. The analysis algorithm is realized in LabVIEW and the spectra of charge, amplitude, signal width and rising time are analyzed offline. The results from Charge-to-Digital Converter, Time-to-Digital Converter and waveform sampling are discussed in detailed comparison.
Use of Electronic Hand-held Devices for Collection of Savannah River Site Environmental Data - 13329
DOE Office of Scientific and Technical Information (OSTI.GOV)
Marberry, Hugh; Moore, Winston
2013-07-01
Savannah River Nuclear Solutions has begun using Xplore Tablet PC's to collect data in the field for soil samples, groundwater samples, air samples and round sheets at the Savannah River Site (SRS). EPA guidelines for groundwater sampling are incorporated into the application to ensure the sample technician follows the proper protocol. The sample technician is guided through the process for sampling and round sheet data collection by a series of menus and input boxes. Field measurements and well stabilization information are entered into the tablet for uploading into Environmental Restoration Data Management System (ERDMS). The process helps to eliminate inputmore » errors and provides data integrity. A soil sample technician has the ability to collect information about location of sample, field parameter, describe the soil sample, print bottle labels, and print chain of custody for the sample that they have collected. An air sample technician has the ability to provide flow, pressure, hours of operation, print bottle labels and chain of custody for samples they collect. Round sheets are collected using the information provided in the various procedures. The data are collected and uploaded into ERDMS. The equipment used is weather proof and hardened for the field use. Global Positioning System (GPS) capabilities are integrated into the applications to provide the location where samples were collected and to help sample technicians locate wells that are not visited often. (authors)« less
Mineralogy and Elemental Composition of Wind Drift Soil at Rocknest, Gale Crater
NASA Technical Reports Server (NTRS)
Blake, D. F.; Bish, D. L.; Morris, R. V.; Downs, R. T.; Trieman, A. H.; Morrison, S. M.; Chipera, S. J.; Ming, D. W.; Yen, A. S.; Vaniman, D. T.;
2013-01-01
The Mars Science Laboratory rover Curiosity has been exploring Mars since August 5, 2012, conducting engineering and first-time activities with its mobility system, arm, sample acquisition and processing system (SA/SPaH-CHIMRA) and science instruments. Curiosity spent 54 sols at a location named "Rocknest," collecting and processing five scoops of loose, unconsolidated materials ("soil") acquired from an aeolian bedform (Fig. 1). The Chemistry and Mineralogy (CheMin) instrument analyzed portions of scoops 3, 4, and 5, to obtain the first quantitative mineralogical analysis of Mars soil, and to provide context for Sample Analysis at Mars (SAM) measurements of volatiles, isotopes and possible organic materials.
One-dimensional ion-beam figuring for grazing-incidence reflective optics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhou, Lin; Idir, Mourad; Bouet, Nathalie
2016-01-01
One-dimensional ion-beam figuring (1D-IBF) can improve grazing-incidence reflective optics, such as Kirkpatrick–Baez mirrors. 1D-IBF requires only one motion degree of freedom, which reduces equipment complexity, resulting in compact and low-cost IBF instrumentation. Furthermore, 1D-IBF is easy to integrate into a single vacuum system with other fabrication processes, such as a thin-film deposition. The NSLS-II Optical Metrology and Fabrication Group has recently integrated the 1D-IBF function into an existing thin-film deposition system by adding an RF ion source to the system. Using a rectangular grid, a 1D removal function needed to perform 1D-IBF has been produced. In this paper, demonstration experimentsmore » of the 1D-IBF process are presented on one spherical and two plane samples. The final residual errors on both plane samples are less than 1 nm r.m.s. In conclusion, the surface error on the spherical sample has been successfully reduced by a factor of 12. The results show that the 1D-IBF method is an effective method to process high-precision 1D synchrotron optics.« less
Floating Ultrasonic Transducer Inspection System and Method for Nondestructive Evaluation
NASA Technical Reports Server (NTRS)
Johnston, Patrick H. (Inventor); Zalameda, Joseph N. (Inventor)
2016-01-01
A method for inspecting a structural sample using ultrasonic energy includes positioning an ultrasonic transducer adjacent to a surface of the sample, and then transmitting ultrasonic energy into the sample. Force pulses are applied to the transducer concurrently with transmission of the ultrasonic energy. A host machine processes ultrasonic return pulses from an ultrasonic pulser/receiver to quantify attenuation of the ultrasonic energy within the sample. The host machine detects a defect in the sample using the quantified level of attenuation. The method may include positioning a dry couplant between an ultrasonic transducer and the surface. A system includes an actuator, an ultrasonic transducer, a dry couplant between the transducer the sample, a scanning device that moves the actuator and transducer, and a measurement system having a pulsed actuator power supply, an ultrasonic pulser/receiver, and a host machine that executes the above method.
NASA Technical Reports Server (NTRS)
Williams, G., Jr.
1983-01-01
The continuous flow electrophoresis system makes electrophoresis possible in a free flowing film of aqueous electrolyte medium. The sample is introduced at one end of the chamber and is subjected to a lateral dc field. This process separates the sample into fractions since each component has a distinctive electrophoric mobility. Evaluations were made of sample conductivity and buffer conductivity as they affect sample band spread and separation using the Continuous Particle Electrophoresis (CPE) system. Samples were prepared from mixtures of 5 percent and 1 percent polystyrene latex (PSL) microspheres which were .4, .56 and .7 microns in diameter. These were prepared in electrolyte media 1x and 3x the conductivity of the curtain buffer, approximately 150 and 450 micro mhos/cm. Samples with matched conductivities produced greater resolution and less band spread than those with 3x the conductivity of the curtain buffer.
Hawker, Charles D; McCarthy, William; Cleveland, David; Messinger, Bonnie L
2014-03-01
Mislabeled samples are a serious problem in most clinical laboratories. Published error rates range from 0.39/1000 to as high as 1.12%. Standardization of bar codes and label formats has not yet achieved the needed improvement. The mislabel rate in our laboratory, although low compared with published rates, prompted us to seek a solution to achieve zero errors. To reduce or eliminate our mislabeled samples, we invented an automated device using 4 cameras to photograph the outside of a sample tube. The system uses optical character recognition (OCR) to look for discrepancies between the patient name in our laboratory information system (LIS) vs the patient name on the customer label. All discrepancies detected by the system's software then require human inspection. The system was installed on our automated track and validated with production samples. We obtained 1 009 830 images during the validation period, and every image was reviewed. OCR passed approximately 75% of the samples, and no mislabeled samples were passed. The 25% failed by the system included 121 samples actually mislabeled by patient name and 148 samples with spelling discrepancies between the patient name on the customer label and the patient name in our LIS. Only 71 of the 121 mislabeled samples detected by OCR were found through our normal quality assurance process. We have invented an automated camera system that uses OCR technology to identify potential mislabeled samples. We have validated this system using samples transported on our automated track. Full implementation of this technology offers the possibility of zero mislabeled samples in the preanalytic stage.
Performance of a segmented HPGe detector at KRISS.
Han, Jubong; Lee, K B; Lee, Jong-Man; Lee, S H; Park, Tae Soon; Oh, J S
2018-04-01
A 24 segmented HPGe coaxial detector was set up with a digitized data acquisition system (DAQ). The DAQ was composed of a digitizer (5 × 10 7 sampling/s), a Field-Programmable Gate Array (FPGA), and a real time operating system. The Full Width Half Maximum (FWHM), rise time, signal characteristics, and spectra of a 137 Cs source were evaluated. The data were processed using an in-house developed gamma-ray tracking system. Copyright © 2017 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Noyes, Ben F.; Mokaberi, Babak; Mandoy, Ram; Pate, Alex; Huijgen, Ralph; McBurney, Mike; Chen, Owen
2017-03-01
Reducing overlay error via an accurate APC feedback system is one of the main challenges in high volume production of the current and future nodes in the semiconductor industry. The overlay feedback system directly affects the number of dies meeting overlay specification and the number of layers requiring dedicated exposure tools through the fabrication flow. Increasing the former number and reducing the latter number is beneficial for the overall efficiency and yield of the fabrication process. An overlay feedback system requires accurate determination of the overlay error, or fingerprint, on exposed wafers in order to determine corrections to be automatically and dynamically applied to the exposure of future wafers. Since current and future nodes require correction per exposure (CPE), the resolution of the overlay fingerprint must be high enough to accommodate CPE in the overlay feedback system, or overlay control module (OCM). Determining a high resolution fingerprint from measured data requires extremely dense overlay sampling that takes a significant amount of measurement time. For static corrections this is acceptable, but in an automated dynamic correction system this method creates extreme bottlenecks for the throughput of said system as new lots have to wait until the previous lot is measured. One solution is using a less dense overlay sampling scheme and employing computationally up-sampled data to a dense fingerprint. That method uses a global fingerprint model over the entire wafer; measured localized overlay errors are therefore not always represented in its up-sampled output. This paper will discuss a hybrid system shown in Fig. 1 that combines a computationally up-sampled fingerprint with the measured data to more accurately capture the actual fingerprint, including local overlay errors. Such a hybrid system is shown to result in reduced modelled residuals while determining the fingerprint, and better on-product overlay performance.
Mercury Deposition Network Site Operator Training for the System Blank and Blind Audit Programs
Wetherbee, Gregory A.; Lehmann, Christopher M.B.
2008-01-01
The U.S. Geological Survey operates the external quality assurance project for the National Atmospheric Deposition Program/Mercury Deposition Network. The project includes the system blank and blind audit programs for assessment of total mercury concentration data quality for wet-deposition samples. This presentation was prepared to train new site operators and to refresh experienced site operators to successfully process and submit system blank and blind audit samples for chemical analysis. Analytical results are used to estimate chemical stability and contamination levels of National Atmospheric Deposition Program/Mercury Deposition Network samples and to evaluate laboratory variability and bias.
System and method for resolving gamma-ray spectra
Gentile, Charles A.; Perry, Jason; Langish, Stephen W.; Silber, Kenneth; Davis, William M.; Mastrovito, Dana
2010-05-04
A system for identifying radionuclide emissions is described. The system includes at least one processor for processing output signals from a radionuclide detecting device, at least one training algorithm run by the at least one processor for analyzing data derived from at least one set of known sample data from the output signals, at least one classification algorithm derived from the training algorithm for classifying unknown sample data, wherein the at least one training algorithm analyzes the at least one sample data set to derive at least one rule used by said classification algorithm for identifying at least one radionuclide emission detected by the detecting device.
Application of mass spectrometry to process control for polymer material in autoclave curing
NASA Technical Reports Server (NTRS)
Smith, A. C.
1983-01-01
Mass spectrometer analysis of gas samples collected during a cure cycle of polymer materials can be used as a process control technique. This technique is particularly helpful in studying the various types of solvents and resin systems used in the preparation of polymer materials and characterizing the chemical composition of different resin systems and their mechanism of polymerization.
ERIC Educational Resources Information Center
Fjermestad, Krister W.; McLeod, Bryce D.; Heiervang, Einar R.; Havik, Odd E.; Ost, Lars-Goran; Haugland, Bente S. M.
2012-01-01
The aim of this study was to examine the factor structure and psychometric properties of an observer-rated youth alliance measure, the Therapy Process Observational Coding System for Child Psychotherapy-Alliance scale (TPOCS-A). The sample was 52 youth diagnosed with anxiety disorders ("M" age = 12.43, "SD" = 2.23, range = 15;…
A Systems View of Mother-Infant Face-to-Face Communication
ERIC Educational Resources Information Center
Beebe, Beatrice; Messinger, Daniel; Bahrick, Lorraine E.; Margolis, Amy; Buck, Karen A.; Chen, Henian
2016-01-01
Principles of a dynamic, dyadic systems view of mother-infant face-to-face communication, which considers self- and interactive processes in relation to one another, were tested. The process of interaction across time in a large low-risk community sample at infant age 4 months was examined. Split-screen videotape was coded on a 1-s time base for…
Irregular analytical errors in diagnostic testing - a novel concept.
Vogeser, Michael; Seger, Christoph
2018-02-23
In laboratory medicine, routine periodic analyses for internal and external quality control measurements interpreted by statistical methods are mandatory for batch clearance. Data analysis of these process-oriented measurements allows for insight into random analytical variation and systematic calibration bias over time. However, in such a setting, any individual sample is not under individual quality control. The quality control measurements act only at the batch level. Quantitative or qualitative data derived for many effects and interferences associated with an individual diagnostic sample can compromise any analyte. It is obvious that a process for a quality-control-sample-based approach of quality assurance is not sensitive to such errors. To address the potential causes and nature of such analytical interference in individual samples more systematically, we suggest the introduction of a new term called the irregular (individual) analytical error. Practically, this term can be applied in any analytical assay that is traceable to a reference measurement system. For an individual sample an irregular analytical error is defined as an inaccuracy (which is the deviation from a reference measurement procedure result) of a test result that is so high it cannot be explained by measurement uncertainty of the utilized routine assay operating within the accepted limitations of the associated process quality control measurements. The deviation can be defined as the linear combination of the process measurement uncertainty and the method bias for the reference measurement system. Such errors should be coined irregular analytical errors of the individual sample. The measurement result is compromised either by an irregular effect associated with the individual composition (matrix) of the sample or an individual single sample associated processing error in the analytical process. Currently, the availability of reference measurement procedures is still highly limited, but LC-isotope-dilution mass spectrometry methods are increasingly used for pre-market validation of routine diagnostic assays (these tests also involve substantial sets of clinical validation samples). Based on this definition/terminology, we list recognized causes of irregular analytical error as a risk catalog for clinical chemistry in this article. These issues include reproducible individual analytical errors (e.g. caused by anti-reagent antibodies) and non-reproducible, sporadic errors (e.g. errors due to incorrect pipetting volume due to air bubbles in a sample), which can both lead to inaccurate results and risks for patients.
A multi-probe thermophoretic soot sampling system for high-pressure diffusion flames
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vargas, Alex M.; Gülder, Ömer L.
Optical diagnostics and physical probing of the soot processes in high pressure combustion pose challenges that are not faced in atmospheric flames. One of the preferred methods of studying soot in atmospheric flames is in situ thermophoretic sampling followed by transmission electron microscopy imaging and analysis for soot sizing and morphology. The application of this method of sampling to high pressures has been held back by various operational and mechanical problems. In this work, we describe a rotating disk multi-probe thermophoretic soot sampling system, driven by a microstepping stepper motor, fitted into a high-pressure chamber capable of producing sooting laminarmore » diffusion flames up to 100 atm. Innovative aspects of the sampling system design include an easy and precise control of the sampling time down to 2.6 ms, avoidance of the drawbacks of the pneumatic drivers used in conventional thermophoretic sampling systems, and the capability to collect ten consecutive samples in a single experimental run. Proof of principle experiments were performed using this system in a laminar diffusion flame of methane, and primary soot diameter distributions at various pressures up to 10 atm were determined. High-speed images of the flame during thermophoretic sampling were recorded to assess the influence of probe intrusion on the flow field of the flame.« less
Measurements of Shock Effects Recorded by Hayabusa Samples
NASA Technical Reports Server (NTRS)
Zolensky, Michael; Mikouchi, Takashi; Hagiya, Kenji; Ohsumi, Kazumasa; Martinez, James; Komatsu, Mutsumi; Chan, Queenie H-.S.
2015-01-01
We requested and have been approved for 5 Hayabusa samples in order definitively establish the degree of shock experienced by the regolith of asteroid Itokawa, and to devise a bridge between shock determinations by standard light optical petrography, crystal structures as determined by synchrotron X-ray diffraction (SXRD), and degree of crystallinity as determined by electron back-scattered diffraction (EBSD) [1,2]. As of the writing of this abstract we are awaiting the approved samples. We propose measurements of astromaterial crystal structures and regolith processes. The proposed research work will improve our understanding of how small, primitive solar system bodies formed and evolved, and improve understanding of the processes that determine the history and future of habitability of environments on other solar system bodies. The results of the proposed research will directly enrich the ongoing asteroid and comet exploration missions by NASA, JAXA and ESA, and broaden our understanding of the origin and evolution of small bodies in the early solar system, and elucidate the nature of asteroid and comet regolith.
Measurements of Shock Effects Recorded by Itokawa Samples
NASA Technical Reports Server (NTRS)
Zolensky, Michael; Mikouchi, Takashi; Hagiya, Kenji; Ohsumi, Kazumasa; Martinez, James; Komatsu, Mutsumi; Chan, Queenie H-.S.
2016-01-01
We requested and have been approved for 5 Hayabusa samples in order definitively establish the degree of shock experienced by the regolith of asteroid Itokawa, and to devise a bridge between shock determinations by standard light optical petrography, crystal structures as determined by synchrotron X-ray diffraction (SXRD), and degree of crystallinity as determined by electron back-scattered diffraction (EBSD). As of the writing of this abstract we are awaiting the approved samples. We propose measurements of astromaterial crystal structures and regolith processes. The proposed research work will improve our understanding of how small, primitive solar system bodies formed and evolved, and improve understanding of the processes that determine the history and future of habitability of environments on other solar system bodies. The results of the proposed research will directly enrich the ongoing asteroid and comet exploration missions by NASA, JAXA and ESA, and broaden our understanding of the origin and evolution of small bodies in the early solar system, and elucidate the nature of asteroid and comet regolith.
Burtis, C.A.; Johnson, W.F.; Walker, W.A.
1985-08-05
A rotor and disc assembly for use in a centrifugal fast analyzer. The assembly is designed to process multiple samples of whole blood followed by aliquoting of the resultant serum into precisely measured samples for subsequent chemical analysis. The assembly requires minimal operator involvement with no mechanical pipetting. The system comprises: (1) a whole blood sample disc; (2) a serum sample disc; (3) a sample preparation rotor; and (4) an analytical rotor. The blood sample disc and serum sample disc are designed with a plurality of precision bore capillary tubes arranged in a spoked array. Samples of blood are loaded into the blood sample disc by capillary action and centrifugally discharged into cavities of the sample preparation rotor where separation of serum and solids is accomplished. The serum is loaded into the capillaries of the serum sample disc by capillary action and subsequently centrifugally expelled into cuvettes of the analyticaly rotor for conventional methods. 5 figs.
Burtis, Carl A.; Johnson, Wayne F.; Walker, William A.
1988-01-01
A rotor and disc assembly for use in a centrifugal fast analyzer. The assembly is designed to process multiple samples of whole blood followed by aliquoting of the resultant serum into precisely measured samples for subsequent chemical analysis. The assembly requires minimal operator involvement with no mechanical pipetting. The system comprises (1) a whole blood sample disc, (2) a serum sample disc, (3) a sample preparation rotor, and (4) an analytical rotor. The blood sample disc and serum sample disc are designed with a plurality of precision bore capillary tubes arranged in a spoked array. Samples of blood are loaded into the blood sample disc in capillary tubes filled by capillary action and centrifugally discharged into cavities of the sample preparation rotor where separation of serum and solids is accomplished. The serum is loaded into the capillaries of the serum sample disc by capillary action and subsequently centrifugally expelled into cuvettes of the analytical rotor for analysis by conventional methods.
Research on photodiode detector-based spatial transient light detection and processing system
NASA Astrophysics Data System (ADS)
Liu, Meiying; Wang, Hu; Liu, Yang; Zhao, Hui; Nan, Meng
2016-10-01
In order to realize real-time signal identification and processing of spatial transient light, the features and the energy of the captured target light signal are first described and quantitatively calculated. Considering that the transient light signal has random occurrence, a short duration and an evident beginning and ending, a photodiode detector based spatial transient light detection and processing system is proposed and designed in this paper. This system has a large field of view and is used to realize non-imaging energy detection of random, transient and weak point target under complex background of spatial environment. Weak signal extraction under strong background is difficult. In this paper, considering that the background signal changes slowly and the target signal changes quickly, filter is adopted for signal's background subtraction. A variable speed sampling is realized by the way of sampling data points with a gradually increased interval. The two dilemmas that real-time processing of large amount of data and power consumption required by the large amount of data needed to be stored are solved. The test results with self-made simulative signal demonstrate the effectiveness of the design scheme. The practical system could be operated reliably. The detection and processing of the target signal under the strong sunlight background was realized. The results indicate that the system can realize real-time detection of target signal's characteristic waveform and monitor the system working parameters. The prototype design could be used in a variety of engineering applications.
Integrated Optical Information Processing
1988-08-01
applications in optical disk memory systems [91. This device is constructed in a glass /SiO2/Si waveguide. The choice of a Si substrate allows for the...contact mask) were formed in the photoresist deposited on all of the samples, we covered the unwanted gratings on each sample with cover glass slides...processing, let us consider TeO2 (v, = 620 m/s) as a potential substrate for applications requiring large time delays. This con- sideration is despite
Spatially resolved imaging of opto-electrical property variations
Nikiforov, Maxim; Darling, Seth B; Suzer, Ozgun; Guest, Jeffrey; Roelofs, Andreas
2014-09-16
Systems and methods for opto electric properties are provided. A light source illuminates a sample. A reference detector senses light from the light source. A sample detector receives light from the sample. A positioning fixture allows for relative positioning of the sample or the light source with respect to each other. An electrical signal device measures the electrical properties of the sample. The reference detector, sample detector and electrical signal device provide information that may be processed to determine opto-electric properties of the same.
Mandal, Abhishek; Boatz, Jennifer C.; Wheeler, Travis; van der Wel, Patrick C. A.
2017-01-01
A number of recent advances in the field of magic-angle-spinning (MAS) solid-state NMR have enabled its application to a range of biological systems of ever increasing complexity. To retain biological relevance, these samples are increasingly studied in a hydrated state. At the same time, experimental feasibility requires the sample preparation process to attain a high sample concentration within the final MAS rotor. We discuss these considerations, and how they have led to a number of different approaches to MAS NMR sample preparation. We describe our experience of how custom-made (or commercially available) ultracentrifugal devices can facilitate a simple, fast and reliable sample preparation process. A number of groups have since adopted such tools, in some cases to prepare samples for sedimentation-style MAS NMR experiments. Here we argue for a more widespread adoption of their use for routine MAS NMR sample preparation. PMID:28229262
Zhou, Hexi; Li, Xiangkun; Chu, Zhaorui; Zhang, Jie
2016-06-01
Effect of temperature downshifts on process performance and bacterial community dynamics was investigated in a bench-scale hybrid A/O system treating real domestic wastewater. Results showed that the average COD removal in this system reached 90.5%, 89.1% and 90.3% for Run 1 (25 °C), Run 2 (15 °C) and Run 3 (10 °C), respectively, and variations in temperature barely affected the effluent COD concentration. The average removal efficiencies of NH4(+)-N were 98.4%, 97.8%, 95.7%, and that of TN were 77.1%, 61.8%, 72% at 25 °C, 15 °C and 10 °C, respectively. Although the hybrid system was subjected to low temperature, this process effectively removed NH4(+)-N and TN even at 10 °C with the average effluent concentrations of 2.4 mg/L and 14.3 mg/L, respectively. Results from high-throughput sequencing analysis revealed that when the operation temperature decreased from 25 °C to 10 °C, the richness and diversity indexes of the system decreased in the sludge samples, while underwent an increase in the biofilm samples. Furthermore, the major heterotrophic bacteria consisted of Lewinella, Lutimonas, Chitinophaga and Fluviicola at 10 °C, which could be central to effective COD removal at low temperature. Additionally, Azospira, one denitrifying-related genus increased from 0.4% to 4.45% in the biofilm samples, with a stable TN removal in response to temperature downshifts. Nitrosomonas and Nitrospira increased significantly in the biofilm samples, implying that the attached biofilm contributed to more nitrification at low temperature. Copyright © 2016 Elsevier Ltd. All rights reserved.
The application of digital signal processing techniques to a teleoperator radar system
NASA Technical Reports Server (NTRS)
Pujol, A.
1982-01-01
A digital signal processing system was studied for the determination of the spectral frequency distribution of echo signals from a teleoperator radar system. The system consisted of a sample and hold circuit, an analog to digital converter, a digital filter, and a Fast Fourier Transform. The system is interfaced to a 16 bit microprocessor. The microprocessor is programmed to control the complete digital signal processing. The digital filtering and Fast Fourier Transform functions are implemented by a S2815 digital filter/utility peripheral chip and a S2814A Fast Fourier Transform chip. The S2815 initially simulates a low-pass Butterworth filter with later expansion to complete filter circuit (bandpass and highpass) synthesizing.
Laurin, Nancy; DeMoors, Anick; Frégeau, Chantal
2012-09-01
Direct amplification of STR loci from biological samples collected on FTA cards without prior DNA purification was evaluated using Identifiler Direct and PowerPlex 16 HS in conjunction with the use of a high throughput Applied Biosystems 3730 DNA Analyzer. In order to reduce the overall sample processing cost, reduced PCR volumes combined with various FTA disk sizes were tested. Optimized STR profiles were obtained using a 0.53 mm disk size in 10 μL PCR volume for both STR systems. These protocols proved effective in generating high quality profiles on the 3730 DNA Analyzer from both blood and buccal FTA samples. Reproducibility, concordance, robustness, sample stability and profile quality were assessed using a collection of blood and buccal samples on FTA cards from volunteer donors as well as from convicted offenders. The new developed protocols offer enhanced throughput capability and cost effectiveness without compromising the robustness and quality of the STR profiles obtained. These results support the use of these protocols for processing convicted offender samples submitted to the National DNA Data Bank of Canada. Similar protocols could be applied to the processing of casework reference samples or in paternity or family relationship testing. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.
A prototype water quality monitoring system is described which offers almost continuous in situ monitoring. The two-man portable system features: (1) a microprocessor controlled central processing unit which allows preprogrammed sampling schedules and reprogramming in situ; (2) a...
Multifrequency Aperture-Synthesizing Microwave Radiometer System (MFASMR). Volume 2: Appendix
NASA Technical Reports Server (NTRS)
Wiley, C. A.; Chang, M. U.
1981-01-01
A number of topics supporting the systems analysis of a multifrequency aperture-synthesizing microwave radiometer system are discussed. Fellgett's (multiple) advantage, interferometer mapping behavior, mapping geometry, image processing programs, and sampling errors are among the topics discussed. A FORTRAN program code is given.
The application of charge-coupled device processors in automatic-control systems
NASA Technical Reports Server (NTRS)
Mcvey, E. S.; Parrish, E. A., Jr.
1977-01-01
The application of charge-coupled device (CCD) processors to automatic-control systems is suggested. CCD processors are a new form of semiconductor component with the unique ability to process sampled signals on an analog basis. Specific implementations of controllers are suggested for linear time-invariant, time-varying, and nonlinear systems. Typical processing time should be only a few microseconds. This form of technology may become competitive with microprocessors and minicomputers in addition to supplementing them.
NASA Astrophysics Data System (ADS)
Yin, Bo; Liu, Li; Wang, Jiahan; Li, Xiran; Liu, Zhenbo; Li, Dewei; Wang, Jun; Liu, Lu; Wu, Jun; Xu, Tingting; Cui, He
2017-10-01
Electric energy measurement as a basic work, an accurate measurements play a vital role for the economic interests of both parties of power supply, the standardized management of the measurement laboratory at all levels is a direct factor that directly affects the fairness of measurement. Currently, the management of metering laboratories generally uses one-dimensional bar code as the recognition object, advances the testing process by manual management, most of the test data requires human input to generate reports. There are many problems and potential risks in this process: Data cannot be saved completely, cannot trace the status of inspection, the inspection process isn't completely controllable and so on. For the provincial metrology center's actual requirements of the whole process management for the performance test of the power measuring appliances, using of large-capacity RF tags as a process management information media, we developed a set of general measurement experiment management system, formulated a standardized full performance test process, improved the raw data recording mode of experimental process, developed a storehouse automatic inventory device, established a strict test sample transfer and storage system, ensured that all the raw data of the inspection can be traced back, achieved full life-cycle control of the sample, significantly improved the quality control level and the effectiveness of inspection work.
Accelerating root system phenotyping of seedlings through a computer-assisted processing pipeline.
Dupuy, Lionel X; Wright, Gladys; Thompson, Jacqueline A; Taylor, Anna; Dekeyser, Sebastien; White, Christopher P; Thomas, William T B; Nightingale, Mark; Hammond, John P; Graham, Neil S; Thomas, Catherine L; Broadley, Martin R; White, Philip J
2017-01-01
There are numerous systems and techniques to measure the growth of plant roots. However, phenotyping large numbers of plant roots for breeding and genetic analyses remains challenging. One major difficulty is to achieve high throughput and resolution at a reasonable cost per plant sample. Here we describe a cost-effective root phenotyping pipeline, on which we perform time and accuracy benchmarking to identify bottlenecks in such pipelines and strategies for their acceleration. Our root phenotyping pipeline was assembled with custom software and low cost material and equipment. Results show that sample preparation and handling of samples during screening are the most time consuming task in root phenotyping. Algorithms can be used to speed up the extraction of root traits from image data, but when applied to large numbers of images, there is a trade-off between time of processing the data and errors contained in the database. Scaling-up root phenotyping to large numbers of genotypes will require not only automation of sample preparation and sample handling, but also efficient algorithms for error detection for more reliable replacement of manual interventions.
Synchrotron/crystal sample preparation
NASA Technical Reports Server (NTRS)
Johnson, R. Barry
1993-01-01
The Center for Applied Optics (CAO) of the University of Alabama in Huntsville (UAH) prepared this final report entitled 'Synchrotron/Crystal Sample Preparation' in completion of contract NAS8-38609, Delivery Order No. 53. Hughes Danbury Optical Systems (HDOS) is manufacturing the Advanced X-ray Astrophysics Facility (AXAF) mirrors. These thin-walled, grazing incidence, Wolter Type-1 mirrors, varying in diameter from 1.2 to 0.68 meters, must be ground and polished using state-of-the-art techniques in order to prevent undue stress due to damage or the presence of crystals and inclusions. The effect of crystals on the polishing and grinding process must also be understood. This involves coating special samples of Zerodur and measuring the reflectivity of the coatings in a synchrotron system. In order to gain the understanding needed on the effect of the Zerodur crystals by the grinding and polishing process, UAH prepared glass samples by cutting, grinding, etching, and polishing as required to meet specifications for witness bars for synchrotron measurements and for investigations of crystals embedded in Zerodur. UAH then characterized these samples for subsurface damage and surface roughness and figure.
Monitoring of beer fermentation based on hybrid electronic tongue.
Kutyła-Olesiuk, Anna; Zaborowski, Michał; Prokaryn, Piotr; Ciosek, Patrycja
2012-10-01
Monitoring of biotechnological processes, including fermentation is extremely important because of the rapidly occurring changes in the composition of the samples during the production. In the case of beer, the analysis of physicochemical parameters allows for the determination of the stage of fermentation process and the control of its possible perturbations. As a tool to control the beer production process a sensor array can be used, composed of potentiometric and voltammetric sensors (so-called hybrid Electronic Tongue, h-ET). The aim of this study is to apply electronic tongue system to distinguish samples obtained during alcoholic fermentation. The samples originate from batch of homemade beer fermentation and from two stages of the process: fermentation reaction and maturation of beer. The applied sensor array consists of 10 miniaturized ion-selective electrodes (potentiometric ET) and silicon based 3-electrode voltammetric transducers (voltammetric ET). The obtained results were processed using Partial Least Squares (PLS) and Partial Least Squares-Discriminant Analysis (PLS-DA). For potentiometric data, voltammetric data, and combined potentiometric and voltammetric data, comparison of the classification ability was conducted based on Root Mean Squared Error (RMSE), sensitivity, specificity, and coefficient F calculation. It is shown, that in the contrast to the separately used techniques, the developed hybrid system allowed for a better characterization of the beer samples. Data fusion in hybrid ET enables to obtain better results both in qualitative analysis (RMSE, specificity, sensitivity) and in quantitative analysis (RMSE, R(2), a, b). Copyright © 2012 Elsevier B.V. All rights reserved.
Data processing, multi-omic pathway mapping, and metabolite activity analysis using XCMS Online
Forsberg, Erica M; Huan, Tao; Rinehart, Duane; Benton, H Paul; Warth, Benedikt; Hilmers, Brian; Siuzdak, Gary
2018-01-01
Systems biology is the study of complex living organisms, and as such, analysis on a systems-wide scale involves the collection of information-dense data sets that are representative of an entire phenotype. To uncover dynamic biological mechanisms, bioinformatics tools have become essential to facilitating data interpretation in large-scale analyses. Global metabolomics is one such method for performing systems biology, as metabolites represent the downstream functional products of ongoing biological processes. We have developed XCMS Online, a platform that enables online metabolomics data processing and interpretation. A systems biology workflow recently implemented within XCMS Online enables rapid metabolic pathway mapping using raw metabolomics data for investigating dysregulated metabolic processes. In addition, this platform supports integration of multi-omic (such as genomic and proteomic) data to garner further systems-wide mechanistic insight. Here, we provide an in-depth procedure showing how to effectively navigate and use the systems biology workflow within XCMS Online without a priori knowledge of the platform, including uploading liquid chromatography (LCLC)–mass spectrometry (MS) data from metabolite-extracted biological samples, defining the job parameters to identify features, correcting for retention time deviations, conducting statistical analysis of features between sample classes and performing predictive metabolic pathway analysis. Additional multi-omics data can be uploaded and overlaid with previously identified pathways to enhance systems-wide analysis of the observed dysregulations. We also describe unique visualization tools to assist in elucidation of statistically significant dysregulated metabolic pathways. Parameter input takes 5–10 min, depending on user experience; data processing typically takes 1–3 h, and data analysis takes ~30 min. PMID:29494574
Insect pest management for raw commodities during storage
USDA-ARS?s Scientific Manuscript database
This book chapter provides an overview of the pest management decision-making process during grain storage. An in-depth discussion of sampling methods, cost-benefit analysis, expert systems, consultants and the use of computer simulation models is provided. Sampling is essential to determine if pest...
NASA Astrophysics Data System (ADS)
Huang, Po-Jung; Baghbani Kordmahale, Sina; Chou, Chao-Kai; Yamaguchi, Hirohito; Hung, Mien-Chie; Kameoka, Jun
2016-03-01
Signal transductions including multiple protein post-translational modifications (PTM), protein-protein interactions (PPI), and protein-nucleic acid interaction (PNI) play critical roles for cell proliferation and differentiation that are directly related to the cancer biology. Traditional methods, like mass spectrometry, immunoprecipitation, fluorescence resonance energy transfer, and fluorescence correlation spectroscopy require a large amount of sample and long processing time. "microchannel for multiple-parameter analysis of proteins in single-complex (mMAPS)"we proposed can reduce the process time and sample volume because this system is composed by microfluidic channels, fluorescence microscopy, and computerized data analysis. In this paper, we will present an automated mMAPS including integrated microfluidic device, automated stage and electrical relay for high-throughput clinical screening. Based on this result, we estimated that this automated detection system will be able to screen approximately 150 patient samples in a 24-hour period, providing a practical application to analyze tissue samples in a clinical setting.
Input-output characterization of an ultrasonic testing system by digital signal analysis
NASA Technical Reports Server (NTRS)
Williams, J. H., Jr.; Lee, S. S.; Karagulle, H.
1986-01-01
Ultrasonic test system input-output characteristics were investigated by directly coupling the transmitting and receiving transducers face to face without a test specimen. Some of the fundamentals of digital signal processing were summarized. Input and output signals were digitized by using a digital oscilloscope, and the digitized data were processed in a microcomputer by using digital signal-processing techniques. The continuous-time test system was modeled as a discrete-time, linear, shift-invariant system. In estimating the unit-sample response and frequency response of the discrete-time system, it was necessary to use digital filtering to remove low-amplitude noise, which interfered with deconvolution calculations. A digital bandpass filter constructed with the assistance of a Blackman window and a rectangular time window were used. Approximations of the impulse response and the frequency response of the continuous-time test system were obtained by linearly interpolating the defining points of the unit-sample response and the frequency response of the discrete-time system. The test system behaved as a linear-phase bandpass filter in the frequency range 0.6 to 2.3 MHz. These frequencies were selected in accordance with the criterion that they were 6 dB below the maximum peak of the amplitude of the frequency response. The output of the system to various inputs was predicted and the results were compared with the corresponding measurements on the system.
Zhang, Rui; Wan, Yi; Ai, Xing; Liu, Zhanqiang; Zhang, Dong
2017-01-01
The physical and chemical properties of bio-titanium alloy implant surfaces play an important role in their corrosion resistance and biological activity. New turning and turning-rolling processes are presented, employing an oxygen-rich environment in order to obtain titanium dioxide layers that can both protect implants from corrosion and also promote cell adhesion. The surface topographies, surface roughnesses and chemical compositions of the sample surfaces were obtained using scanning electron microscopy, a white light interferometer, and the Auger electron spectroscopy, respectively. The corrosion resistance of the samples in a simulated body fluid was determined using electrochemical testing. Biological activity on the samples was also analyzed, using a vitro cell culture system. The results show that compared with titanium oxide layers formed using a turning process in air, the thickness of the titanium oxide layers formed using turning and turning-rolling processes in an oxygen-rich environment increased by 4.6 and 7.3 times, respectively. Using an oxygen-rich atmosphere in the rolling process greatly improves the corrosion resistance of the resulting samples in a simulated body fluid. On samples produced using the turning-rolling process, cells spread quickly and exhibited the best adhesion characteristics.
How Is This Flower Pollinated? A Polyclave Key to Use in Teaching.
ERIC Educational Resources Information Center
Tyrrell, Lucy
1989-01-01
Presents an identification method which uses the process of elimination to identify pollination systems. Provides the polyclave key, methodology for using the key, a sample worksheet, and abbreviation codes for pollination systems. (MVL)
NASA Astrophysics Data System (ADS)
Miller, David P.; Bonaccorsi, Rosalba; Davis, Kiel
2008-10-01
Mars Astrobiology Research and Technology Experiment (MARTE) investigators used an automated drill and sample processing hardware to detect and categorize life-forms found in subsurface rock at Río Tinto, Spain. For the science to be successful, it was necessary for the biomass from other sources -- whether from previously processed samples (cross contamination) or the terrestrial environment (forward contamination) -- to be insignificant. The hardware and practices used in MARTE were designed around this problem. Here, we describe some of the design issues that were faced and classify them into problems that are unique to terrestrial tests versus problems that would also exist for a system that was flown to Mars. Assessment of the biomass at various stages in the sample handling process revealed mixed results; the instrument design seemed to minimize cross contamination, but contamination from the surrounding environment sometimes made its way onto the surface of samples. Techniques used during the MARTE Río Tinto project, such as facing the sample, appear to remove this environmental contamination without introducing significant cross contamination from previous samples.
Miller, David P; Bonaccorsi, Rosalba; Davis, Kiel
2008-10-01
Mars Astrobiology Research and Technology Experiment (MARTE) investigators used an automated drill and sample processing hardware to detect and categorize life-forms found in subsurface rock at Río Tinto, Spain. For the science to be successful, it was necessary for the biomass from other sources--whether from previously processed samples (cross contamination) or the terrestrial environment (forward contamination)-to be insignificant. The hardware and practices used in MARTE were designed around this problem. Here, we describe some of the design issues that were faced and classify them into problems that are unique to terrestrial tests versus problems that would also exist for a system that was flown to Mars. Assessment of the biomass at various stages in the sample handling process revealed mixed results; the instrument design seemed to minimize cross contamination, but contamination from the surrounding environment sometimes made its way onto the surface of samples. Techniques used during the MARTE Río Tinto project, such as facing the sample, appear to remove this environmental contamination without introducing significant cross contamination from previous samples.
Laser-aided material identification for the waste sorting process
NASA Astrophysics Data System (ADS)
Haferkamp, Heinz; Burmester, Ingo; Engel, Kai
1994-03-01
The LZH has carried out investigations in the field of rapid laser-supported material- identification systems for automatic material-sorting systems. The aim of this research is the fast identification of different sorts of plastics coming from recycled rubbish or electronic waste. Within a few milliseconds a spot on the sample which has to be identified is heated with a CO2 laser. The different and specific chemical and physical material properties of the examined sample cause a different temperature distribution on the surface which is measured with an IR thermographic system. This `thermal impulse response' has to be analyzed by means of a computer system. The results of previous investigations have shown that material identification of different sorts of plastics can possibly be done at a frequency of 30 Hz. Due to economic efficiency, a high velocity identification process is necessary to sort huge waste currents.
Timing Recovery Strategies in Magnetic Recording Systems
NASA Astrophysics Data System (ADS)
Kovintavewat, Piya
At some point in a digital communications receiver, the received analog signal must be sampled. Good performance requires that these samples be taken at the right times. The process of synchronizing the sampler with the received analog waveform is known as timing recovery. Conventional timing recovery techniques perform well only when operating at high signal-to-noise ratio (SNR). Nonetheless, iterative error-control codes allow reliable communication at very low SNR, where conventional techniques fail. This paper provides a detailed review on the timing recovery strategies based on per-survivor processing (PSP) that are capable of working at low SNR. We also investigate their performance in magnetic recording systems because magnetic recording is a primary method of storage for a variety of applications, including desktop, mobile, and server systems. Results indicate that the timing recovery strategies based on PSP perform better than the conventional ones and are thus worth being employed in magnetic recording systems.
A low cost Doppler system for vascular dialysis access surveillance.
Molina, P S C; Moraes, R; Baggio, J F R; Tognon, E A
2004-01-01
The National Kidney Foundation guidelines for vascular access recommend access surveillance to avoid morbidity among patients undergoing hemodialysis. Methods to detect access failure based on CW Doppler system are being proposed to implement surveillance programs at lower cost. This work describes a low cost Doppler system implemented in a PC notebook designed to carry out this task. A Doppler board samples the blood flow velocity and delivers demodulated quadrature Doppler signals. These signals are sampled by a notebook sound card. Software for Windows OS (running at the notebook) applies CFFT to consecutive 11.6 ms intervals of Doppler signals. The sonogram is presented on the screen in real time. The software also calculates the maximum and the intensity weighted mean frequency envelopes. Since similar systems employ DSP boards to process the Doppler signals, cost reduction was achieved. The Doppler board electronic circuits and routines to process the Doppler signals are presented.
System for sensing droplet formation time delay in a flow cytometer
Van den Engh, Ger; Esposito, Richard J.
1997-01-01
A droplet flow cytometer system which includes a system to optimize the droplet formation time delay based on conditions actually experienced includes an automatic droplet sampler which rapidly moves a plurality of containers stepwise through the droplet stream while simultaneously adjusting the droplet time delay. Through the system sampling of an actual substance to be processed can be used to minimize the effect of the substances variations or the determination of which time delay is optimal. Analysis such as cell counting and the like may be conducted manually or automatically and input to a time delay adjustment which may then act with analysis equipment to revise the time delay estimate actually applied during processing. The automatic sampler can be controlled through a microprocessor and appropriate programming to bracket an initial droplet formation time delay estimate. When maximization counts through volume, weight, or other types of analysis exists in the containers, the increment may then be reduced for a more accurate ultimate setting. This may be accomplished while actually processing the sample without interruption.
TQM (Total Quality Management) SPARC (Special Process Action Review Committees) Handbook
1989-08-01
This document describes the techniques used to support and guide the Special Process Action Review Committees for accomplishing their goals for Total Quality Management (TQM). It includes concepts and definitions, checklists, sample formats, and assessment criteria. Keywords: Continuous process improvement; Logistics information; Process analysis; Quality control; Quality assurance; Total Quality Management ; Statistical processes; Management Planning and control; Management training; Management information systems.
Legacy sample disposition project. Volume 2: Final report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gurley, R.N.; Shifty, K.L.
1998-02-01
This report describes the legacy sample disposition project at the Idaho Engineering and Environmental Laboratory (INEEL), which assessed Site-wide facilities/areas to locate legacy samples and owner organizations and then characterized and dispositioned these samples. This project resulted from an Idaho Department of Environmental Quality inspection of selected areas of the INEEL in January 1996, which identified some samples at the Test Reactor Area and Idaho Chemical Processing Plant that had not been characterized and dispositioned according to Resource Conservation and Recovery Act (RCRA) requirements. The objective of the project was to manage legacy samples in accordance with all applicable environmentalmore » and safety requirements. A systems engineering approach was used throughout the project, which included collecting the legacy sample information and developing a system for amending and retrieving the information. All legacy samples were dispositioned by the end of 1997. Closure of the legacy sample issue was achieved through these actions.« less
Vibration Pattern Imager (VPI): A control and data acquisition system for scanning laser vibrometers
NASA Technical Reports Server (NTRS)
Rizzi, Stephen A.; Brown, Donald E.; Shaffer, Thomas A.
1993-01-01
The Vibration Pattern Imager (VPI) system was designed to control and acquire data from scanning laser vibrometer sensors. The PC computer based system uses a digital signal processing (DSP) board and an analog I/O board to control the sensor and to process the data. The VPI system was originally developed for use with the Ometron VPI Sensor, but can be readily adapted to any commercially available sensor which provides an analog output signal and requires analog inputs for control of mirror positioning. The sensor itself is not part of the VPI system. A graphical interface program, which runs on a PC under the MS-DOS operating system, functions in an interactive mode and communicates with the DSP and I/O boards in a user-friendly fashion through the aid of pop-up menus. Two types of data may be acquired with the VPI system: single point or 'full field.' In the single point mode, time series data is sampled by the A/D converter on the I/O board (at a user-defined sampling rate for a selectable number of samples) and is stored by the PC. The position of the measuring point (adjusted by mirrors in the sensor) is controlled via a mouse input. The mouse input is translated to output voltages by the D/A converter on the I/O board to control the mirror servos. In the 'full field' mode, the measurement point is moved over a user-selectable rectangular area. The time series data is sampled by the A/D converter on the I/O board (at a user-defined sampling rate for a selectable number of samples) and converted to a root-mean-square (rms) value by the DSP board. The rms 'full field' velocity distribution is then uploaded for display and storage on the PC.
A novel rotating experimental platform in a superconducting magnet.
Chen, Da; Cao, Hui-Ling; Ye, Ya-Jing; Dong, Chen; Liu, Yong-Ming; Shang, Peng; Yin, Da-Chuan
2016-08-01
This paper introduces a novel platform designed to be used in a strong static magnetic field (in a superconducting magnet). The platform is a sample holder that rotates in the strong magnetic field. Any samples placed in the platform will rotate due to the rotation of the sample holder. With this platform, a number of experiments such as material processing, culture of biological systems, chemical reactions, or other processes can be carried out. In this report, we present some preliminary experiments (protein crystallization, cell culture, and seed germination) conducted using this platform. The experimental results showed that the platform can affect the processes, indicating that it provides a novel environment that has not been investigated before and that the effects of such an environment on many different physical, chemical, or biological processes can be potentially useful for applications in many fields.
Code of Federal Regulations, 2010 CFR
2010-01-01
... management of product, process or service quality and environmental performance. Sampling means the selection..., process, service, or person's qualifications conforms to specified requirements. Conformity assessment.... Requirements for products, services, systems, and organizations are those defined by law or regulation or by an...
Material Science Experiments on Mir
NASA Technical Reports Server (NTRS)
Kroes, Roger L.
1999-01-01
This paper describes the microgravity materials experiments carried out on the Shuttle/Mir program. There were six experiments, all of which investigated some aspect of diffusivity in liquid melts. The Liquid Metal Diffusion (LMD) experiment investigated the diffusivity of molten Indium samples at 185 C using a radioactive tracer, In-114m. By monitoring two different gamma ray energies (190 keV and 24 keV) emitted by the samples it was possible to measure independently the diffusion rates in the bulk and at the surface of the samples. The Queens University Experiment in Liquid Diffusion (QUELD) was the furnace facility used to process 213 samples for the five other experiments. These experiments investigated the diffusion, ripening, crystal growth, and glass formation in metal, semiconductor, and glass samples. This facility had the capability to process samples in an isothermal or gradient configuration for varying periods of time at temperatures up to 900 C. Both the LMD and the QUELD furnaces were mounted on the Microgravity Isolation Mount (MIM) which provided isolation from g-jitter. All the microgravity experiments were supported by the Space Acceleration Measurement System (SAMS); a three head three axes acceleration monitoring system which measured and recorded the acceleration environment.
NASA Technical Reports Server (NTRS)
1982-01-01
The basic test methods of aging and deterioration mechanisms of electrical insulating materials are discussed. A comprehensive test system developed to study the degradation process is described. This system is completely checked, and calibrated with a few insulating material samples.
Internship Progress Summary: Fall 2016
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wiser, Ralph S.; Valencia, Matthew John
2016-12-13
This fall I had the opportunity to work at Los Alamos National Laboratory for the Technology Applications engineering group. I assisted two main projects during my appointment, both related to the Lab’s mission statement: “To solve national security challenges through scientific excellence.” My first project, a thermal source transfer unit, involved skills such as mechanical design, heat transfer simulation, and design analysis. The goal was to create a container that could protect a heat source and regulate its temperature during transit. I generated several designs, performed heat transfer simulations, and chose a design for prototyping. The second project was amore » soil drying unit for use in post blast sample analysis. To ensure fast and accurate sample processing, agents in the field wanted a system that could process wet dirt and turn it into dry powder. We designed a system of commercially available parts, and we tested the systems to determine the best methods and processes.« less
Hu, Jian Zhi [Richland, WA; Sears, Jr., Jesse A.; Hoyt, David W [Richland, WA; Wind, Robert A [Kennewick, WA
2009-05-19
Described are a "Discrete Magic Angle Turning" (DMAT) system, devices, and processes that combine advantages of both magic angle turning (MAT) and magic angle hopping (MAH) suitable, e.g., for in situ magnetic resonance spectroscopy and/or imaging. In an exemplary system, device, and process, samples are rotated in a clockwise direction followed by an anticlockwise direction of exactly the same amount. Rotation proceeds through an angle that is typically greater than about 240 degrees but less than or equal to about 360 degrees at constant speed for a time applicable to the evolution dimension. Back and forth rotation can be synchronized and repeated with a special radio frequency (RF) pulse sequence to produce an isotropic-anisotropic shift 2D correlation spectrum. The design permits tubes to be inserted into the sample container without introducing plumbing interferences, further allowing control over such conditions as temperature, pressure, flow conditions, and feed compositions, thus permitting true in-situ investigations to be carried out.
Minimally processed vegetable salads: microbial quality evaluation.
Fröder, Hans; Martins, Cecília Geraldes; De Souza, Katia Leani Oliveira; Landgraf, Mariza; Franco, Bernadette D G M; Destro, Maria Teresa
2007-05-01
The increasing demand for fresh fruits and vegetables and for convenience foods is causing an expansion of the market share for minimally processed vegetables. Among the more common pathogenic microorganisms that can be transmitted to humans by these products are Listeria monocytogenes, Escherichia coli O157:H7, and Salmonella. The aim of this study was to evaluate the microbial quality of a selection of minimally processed vegetables. A total of 181 samples of minimally processed leafy salads were collected from retailers in the city of Sao Paulo, Brazil. Counts of total coliforms, fecal coliforms, Enterobacteriaceae, psychrotrophic microorganisms, and Salmonella were conducted for 133 samples. L. monocytogenes was assessed in 181 samples using the BAX System and by plating the enrichment broth onto Palcam and Oxford agars. Suspected Listeria colonies were submitted to classical biochemical tests. Populations of psychrotrophic microorganisms >10(6) CFU/g were found in 51% of the 133 samples, and Enterobacteriaceae populations between 10(5) and 106 CFU/g were found in 42% of the samples. Fecal coliform concentrations higher than 10(2) CFU/g (Brazilian standard) were found in 97 (73%) of the samples, and Salmonella was detected in 4 (3%) of the samples. Two of the Salmonella-positive samples had <10(2) CFU/g concentrations of fecal coliforms. L. monocytogenes was detected in only 1 (0.6%) of the 181 samples examined. This positive sample was simultaneously detected by both methods. The other Listeria species identified by plating were L. welshimeri (one sample of curly lettuce) and L. innocua (2 samples of watercress). The results indicate that minimally processed vegetables had poor microbiological quality, and these products could be a vehicle for pathogens such as Salmonella and L. monocytogenes.
Lab-on-a-chip based total-phosphorus analysis device utilizing a photocatalytic reaction
NASA Astrophysics Data System (ADS)
Jung, Dong Geon; Jung, Daewoong; Kong, Seong Ho
2018-02-01
A lab-on-a-chip (LOC) device for total phosphorus (TP) analysis was fabricated for water quality monitoring. Many commercially available TP analysis systems used to estimate water quality have good sensitivity and accuracy. However, these systems also have many disadvantages such as bulky size, complex pretreatment processes, and high cost, which limit their application. In particular, conventional TP analysis systems require an indispensable pretreatment step, in which the fluidic analyte is heated to 120 °C for 30 min to release the dissolved phosphate, because many phosphates are soluble in water at a standard temperature and pressure. In addition, this pretreatment process requires elevated pressures of up to 1.1 kg cm-2 in order to prevent the evaporation of the heated analyte. Because of these limiting conditions required by the pretreatment processes used in conventional systems, it is difficult to miniaturize TP analysis systems. In this study, we employed a photocatalytic reaction in the pretreatment process. The reaction was carried out by illuminating a photocatalytic titanium dioxide (TiO2) surface formed in a microfluidic channel with ultraviolet (UV) light. This pretreatment process does not require elevated temperatures and pressures. By applying this simplified, photocatalytic-reaction-based pretreatment process to a TP analysis system, greater degrees of freedom are conferred to the design and fabrication of LOC devices for TP monitoring. The fabricated LOC device presented in this paper was characterized by measuring the TP concentration of an unknown sample, and comparing the results with those measured by a conventional TP analysis system. The TP concentrations of the unknown sample measured by the proposed LOC device and the conventional TP analysis system were 0.018 mgP/25 mL and 0.019 mgP/25 mL, respectively. The experimental results revealed that the proposed LOC device had a performance comparable to the conventional bulky TP analysis system. Therefore, our device could be directly employed in water quality monitoring as an alternative to conventional TP analysis systems.
Vargas, E; Ruiz, M A; Campuzano, S; Reviejo, A J; Pingarrón, J M
2016-03-31
A non-destructive, rapid and simple to use sensing method for direct determination of glucose in non-processed fruits is described. The strategy involved on-line microdialysis sampling coupled with a continuous flow system with amperometric detection at an enzymatic biosensor. Apart from direct determination of glucose in fruit juices and blended fruits, this work describes for the first time the successful application of an enzymatic biosensor-based electrochemical approach to the non-invasive determination of glucose in raw fruits. The methodology correlates, through previous calibration set-up, the amperometric signal generated from glucose in non-processed fruits with its content in % (w/w). The comparison of the obtained results using the proposed approach in different fruits with those provided by other method involving the same commercial biosensor as amperometric detector in stirred solutions pointed out that there were no significant differences. Moreover, in comparison with other available methodologies, this microdialysis-coupled continuous flow system amperometric biosensor-based procedure features straightforward sample preparation, low cost, reduced assay time (sampling rate of 7 h(-1)) and ease of automation. Copyright © 2016 Elsevier B.V. All rights reserved.
ISPyB: an information management system for synchrotron macromolecular crystallography.
Delagenière, Solange; Brenchereau, Patrice; Launer, Ludovic; Ashton, Alun W; Leal, Ricardo; Veyrier, Stéphanie; Gabadinho, José; Gordon, Elspeth J; Jones, Samuel D; Levik, Karl Erik; McSweeney, Seán M; Monaco, Stéphanie; Nanao, Max; Spruce, Darren; Svensson, Olof; Walsh, Martin A; Leonard, Gordon A
2011-11-15
Individual research groups now analyze thousands of samples per year at synchrotron macromolecular crystallography (MX) resources. The efficient management of experimental data is thus essential if the best possible experiments are to be performed and the best possible data used in downstream processes in structure determination pipelines. Information System for Protein crystallography Beamlines (ISPyB), a Laboratory Information Management System (LIMS) with an underlying data model allowing for the integration of analyses down-stream of the data collection experiment was developed to facilitate such data management. ISPyB is now a multisite, generic LIMS for synchrotron-based MX experiments. Its initial functionality has been enhanced to include improved sample tracking and reporting of experimental protocols, the direct ranking of the diffraction characteristics of individual samples and the archiving of raw data and results from ancillary experiments and post-experiment data processing protocols. This latter feature paves the way for ISPyB to play a central role in future macromolecular structure solution pipelines and validates the application of the approach used in ISPyB to other experimental techniques, such as biological solution Small Angle X-ray Scattering and spectroscopy, which have similar sample tracking and data handling requirements.
Fiber-optic temperature probe system for inner body
NASA Astrophysics Data System (ADS)
Liu, Bo; Deng, Xing-Zhong; Cao, Wei; Cheng, Xianping; Xie, Tuqiang; Zhong, Zugen
1991-08-01
The authors have designed a fiber-optic temperature probe system that can quickly insert its probe into bodies to measure temperature. Its thermometer unit has the function of program- controlled zeroing. The single-chip microcomputer is used to control the whole system and process data. The sample system has been tested in a coal furnace.
The planetary spatial data infrastructure for the OSIRIS-REx mission
NASA Astrophysics Data System (ADS)
DellaGiustina, D. N.; Selznick, S.; Nolan, M. C.; Enos, H. L.; Lauretta, D. S.
2017-12-01
The primary objective of the Origins, Spectral Interpretation, Resource Identification, and Security-Regolith Explorer (OSIRIS-REx) mission is to return a pristine sample of carbonaceous material from primitive asteroid (101955) Bennu. Understanding the geospatial context of Bennu is critical to choosing a sample-site and also linking the nature of the sample to the global properties of Bennu and the broader asteroid population. We established a planetary spatial data infrastructure (PSDI) support the primary objective of OSIRIS-REx. OSIRIS-REx is unique among planetary missions in that all remote sensing is performed to support the sample return objective. Prior to sampling, OSIRIS-REx will survey Bennu for nearly two years to select and document the most valuable primary and backup sample sites. During this period, the mission will combine coordinated observations from five science instruments into four thematic maps: deliverability, safety, sampleability, and scientific value. The deliverability map assesses the probability that the flight dynamics team can deliver the spacecraft to the desired location. The safety map indicates the probability that physical hazards are present at the sample-site. The sampleability map quantifies the probability that a sample can be successfully collected from the surface. Finally, the scientific value map shows the probability that the collected sample contains organics and volatiles and also places the sample site in a definitive geological context relative to Bennu's history. The OSIRIS-REx Science Processing and Operations Center (SPOC) serves as the operational PSDI for the mission. The SPOC is tasked with intake of all data from the spacecraft and other ground sources and assimilating these data into a single comprehensive system for processing and presentation. The SPOC centralizes all geographic data of Bennu in a relational database and ensures that standardization and provenance are maintained throughout proximity operations.The SPOC is a live system that handles inputs from spacecraft and science instrument telemetry, and science data producers. It includes multiple levels of validation, both automated and manual to process all data in a robust and reliable manner and eventually deliver it to the NASA Planetary Data System for archive.
Lam, H K
2012-02-01
This paper investigates the stability of sampled-data output-feedback (SDOF) polynomial-fuzzy-model-based control systems. Representing the nonlinear plant using a polynomial fuzzy model, an SDOF fuzzy controller is proposed to perform the control process using the system output information. As only the system output is available for feedback compensation, it is more challenging for the controller design and system analysis compared to the full-state-feedback case. Furthermore, because of the sampling activity, the control signal is kept constant by the zero-order hold during the sampling period, which complicates the system dynamics and makes the stability analysis more difficult. In this paper, two cases of SDOF fuzzy controllers, which either share the same number of fuzzy rules or not, are considered. The system stability is investigated based on the Lyapunov stability theory using the sum-of-squares (SOS) approach. SOS-based stability conditions are obtained to guarantee the system stability and synthesize the SDOF fuzzy controller. Simulation examples are given to demonstrate the merits of the proposed SDOF fuzzy control approach.
NASA Technical Reports Server (NTRS)
Cramer, Christopher J.; Wright, James D.; Simmons, Scott A.; Bobbitt, Lynn E.; DeMoss, Joshua A.
2015-01-01
The paper will present a brief background of the previous data acquisition system at the National Transonic Facility (NTF) and the reasoning and goals behind the upgrade to the current Test SLATE (Test Software Laboratory and Automated Testing Environments) data acquisition system. The components, performance characteristics, and layout of the Test SLATE system within the NTF control room will be discussed. The development, testing, and integration of Test SLATE within NTF operations will be detailed. The operational capabilities of the system will be outlined including: test setup, instrumentation calibration, automatic test sequencer setup, data recording, communication between data and facility control systems, real time display monitoring, and data reduction. The current operational status of the Test SLATE system and its performance during recent NTF testing will be highlighted including high-speed, frame-by-frame data acquisition with conditional sampling post-processing applied. The paper concludes with current development work on the system including the capability for real-time conditional sampling during data acquisition and further efficiency enhancements to the wind tunnel testing process.
USDA-ARS?s Scientific Manuscript database
Using five centimeter resolution images acquired with an unmanned aircraft system (UAS), we developed and evaluated an image processing workflow that included the integration of resolution-appropriate field sampling, feature selection, object-based image analysis, and processing approaches for UAS i...
40 CFR 63.166 - Standards: Sampling connection systems.
Code of Federal Regulations, 2012 CFR
2012-07-01
... fluid to a process; or (3) Be designed and operated to capture and transport the purged process fluid to a control device that complies with the requirements of § 63.172 of this subpart; or (4) Collect... of subpart G of this part applicable to group 1 wastewater streams. If the purged process fluid does...
40 CFR 63.166 - Standards: Sampling connection systems.
Code of Federal Regulations, 2014 CFR
2014-07-01
... fluid to a process; or (3) Be designed and operated to capture and transport the purged process fluid to a control device that complies with the requirements of § 63.172 of this subpart; or (4) Collect... of subpart G of this part applicable to group 1 wastewater streams. If the purged process fluid does...
A Mixed Kijima Model Using the Weibull-Based Generalized Renewal Processes
2015-01-01
Generalized Renewal Processes are useful for approaching the rejuvenation of dynamical systems resulting from planned or unplanned interventions. We present new perspectives for the Generalized Renewal Processes in general and for the Weibull-based Generalized Renewal Processes in particular. Disregarding from literature, we present a mixed Generalized Renewal Processes approach involving Kijima Type I and II models, allowing one to infer the impact of distinct interventions on the performance of the system under study. The first and second theoretical moments of this model are introduced as well as its maximum likelihood estimation and random sampling approaches. In order to illustrate the usefulness of the proposed Weibull-based Generalized Renewal Processes model, some real data sets involving improving, stable, and deteriorating systems are used. PMID:26197222
Dynamic measurement of fluorescent proteins spectral distribution on virus infected cells
NASA Astrophysics Data System (ADS)
Lee, Ja-Yun; Wu, Ming-Xiu; Kao, Chia-Yun; Wu, Tzong-Yuan; Hsu, I.-Jen
2006-09-01
We constructed a dynamic spectroscopy system that can simultaneously measure the intensity and spectral distributions of samples with multi-fluorophores in a single scan. The system was used to monitor the fluorescence distribution of cells infected by the virus, which is constructed by a recombinant baculoviruses, vAcD-Rhir-E, containing the red and green fluorescent protein gene that can simultaneously produce dual fluorescence in recombinant virus-infected Spodoptera frugiperda 21 cells (Sf21) under the control of a polyhedrin promoter. The system was composed of an excitation light source, a scanning system and a spectrometer. We also developed an algorithm and fitting process to analyze the pattern of fluorescence distribution of the dual fluorescence produced in the recombinant virus-infected cells. All the algorithm and calculation are automatically processed in a visualized scanning program and can monitor the specific region of sample by calculating its intensity distribution. The spectral measurement of each pixel was performed at millisecond range and the two dimensional distribution of full spectrum was recorded within several seconds. We have constructed a dynamic spectroscopy system to monitor the process of virus-infection of cells. The distributions of the dual fluorescence were simultaneously measured at micrometer resolution.
X-ray tomographic image magnification process, system and apparatus therefor
Kinney, J.H.; Bonse, U.K.; Johnson, Q.C.; Nichols, M.C.; Saroyan, R.A.; Massey, W.N.; Nusshardt, R.
1993-09-14
A computerized three-dimensional x-ray tomographic microscopy system is disclosed, comprising: (a) source means for providing a source of parallel x-ray beams, (b) staging means for staging and sequentially rotating a sample to be positioned in the path of the (c) x-ray image magnifier means positioned in the path of the beams downstream from the sample, (d) detecting means for detecting the beams after being passed through and magnified by the image magnifier means, and (e) computing means for analyzing values received from the detecting means, and converting the values into three-dimensional representations. Also disclosed is a process for magnifying an x-ray image, and apparatus therefor. 25 figures.
X-ray tomographic image magnification process, system and apparatus therefor
Kinney, John H.; Bonse, Ulrich K.; Johnson, Quintin C.; Nichols, Monte C.; Saroyan, Ralph A.; Massey, Warren N.; Nusshardt, Rudolph
1993-01-01
A computerized three-dimensional x-ray tomographic microscopy system is disclosed, comprising: a) source means for providing a source of parallel x-ray beams, b) staging means for staging and sequentially rotating a sample to be positioned in the path of the c) x-ray image magnifier means positioned in the path of the beams downstream from the sample, d) detecting means for detecting the beams after being passed through and magnified by the image magnifier means, and e) computing means for analyzing values received from the detecting means, and converting the values into three-dimensional representations. Also disclosed is a process for magnifying an x-ray image, and apparatus therefor.
Investigating the Randomness of Numbers
ERIC Educational Resources Information Center
Pendleton, Kenn L.
2009-01-01
The use of random numbers is pervasive in today's world. Random numbers have practical applications in such far-flung arenas as computer simulations, cryptography, gambling, the legal system, statistical sampling, and even the war on terrorism. Evaluating the randomness of extremely large samples is a complex, intricate process. However, the…
Matuszewska, Renata; Szczotko, Maciej; Krogulska, Bozena
2012-01-01
The presence of parasitic protozoa in drinking water is mostly a result of improperly maintened the water treatment process. Currently, in Poland the testing of Cryptosporidium and Giardia in water as a part of routine monitoring of water is not perform. The aim of this study was the optimization of the method of Cryptosporidium and Giardia detection in water according to the main principles of standard ISO 15553:2006 and using Filta-Max xpress automatic elution station. Preliminary tests were performed on the samples contaminated with oocysts and cysts of reference strains of both parasitic protozoa. Further studies were carried out on environmental samples of surface water sampled directly from the intakes of water (21 samples from Vistula River and 8 samples from Zegrzynski Lake). Filtration process and samples volume reducing were performed using an automatic elution system Filta-Max xpress. Next, samples were purified during immunomagnetic separation process (IMS). Isolated cysts and oocysts were stained with FITC and DAPI and than the microscopic observation using an epifluorescence microscope was carried out. Recovery of parasite protozoa in all contaminated water samples after 9-cycles elution process applied was mean 60.6% for Cryptosporidium oocysts and 36.1% for Giardia cysts. Studies on the environmental surface water samples showed the presence of both parasitic protozoa. Number of detected Giardia cysts ranged from 1.0/10 L up to 4.5/10 L in samples from Zegrzynski Lake and from 1.0/10 L up to 38.9/10 L in samples from Vistula River. Cryptosporidium oocysts were present in 50% of samples from the Zegrzynski Lake and in 47.6% of samples from the Vistula River, and their number in both cases was similar and ranged from 0.5 up to 2.5 oocyst/10 L. The results show that applied procedure is appropriate for detection the presence of parasitic protosoan in water, but when water contains much amount of inorganic matter and suspended solids test method have to be modified like subsamples preparation and filtration process speed reduction. The applied method with the modification using Filta-Max xpress system can be useful for the routine monitoring of water. Detection of Cryptosporidium and Giardia in all samples of water taken from the intakes of surface water shows the possibility oftransfering of the protozoan cysts into the water intended for the consumption, therefore the testing of Cryptosporidium and Giardia should be included into the monitoring of water.
Pre-Mission Input Requirements to Enable Successful Sample Collection by A Remote Field/EVA Team
NASA Technical Reports Server (NTRS)
Cohen, B. A.; Lim, D. S. S.; Young, K. E.; Brunner, A.; Elphic, R. E.; Horne, A.; Kerrigan, M. C.; Osinski, G. R.; Skok, J. R.; Squyres, S. W.;
2016-01-01
The FINESSE (Field Investigations to Enable Solar System Science and Exploration) team, part of the Solar System Exploration Virtual Institute (SSERVI), is a field-based research program aimed at generating strategic knowledge in preparation for human and robotic exploration of the Moon, near-Earth asteroids, Phobos and Deimos, and beyond. In contract to other technology-driven NASA analog studies, The FINESSE WCIS activity is science-focused and, moreover, is sampling-focused with the explicit intent to return the best samples for geochronology studies in the laboratory. We used the FINESSE field excursion to the West Clearwater Lake Impact structure (WCIS) as an opportunity to test factors related to sampling decisions. We examined the in situ sample characterization and real-time decision-making process of the astronauts, with a guiding hypothesis that pre-mission training that included detailed background information on the analytical fate of a sample would better enable future astronauts to select samples that would best meet science requirements. We conducted three tests of this hypothesis over several days in the field. Our investigation was designed to document processes, tools and procedures for crew sampling of planetary targets. This was not meant to be a blind, controlled test of crew efficacy, but rather an effort to explicitly recognize the relevant variables that enter into sampling protocol and to be able to develop recommendations for crew and backroom training in future endeavors.
Sutherland, J David; Tu, Noah P; Nemcek, Thomas A; Searle, Philip A; Hochlowski, Jill E; Djuric, Stevan W; Pan, Jeffrey Y
2014-04-01
A flexible and integrated flow-chemistry-synthesis-purification compound-generation and sample-management platform has been developed to accelerate the production of small-molecule organic-compound drug candidates in pharmaceutical research. Central to the integrated system is a Mitsubishi robot, which hands off samples throughout the process to the next station, including synthesis and purification, sample dispensing for purity and quantification analysis, dry-down, and aliquot generation.
NASA Astrophysics Data System (ADS)
Rim, Jung H.
Accurate and fast determination of the activity of radionuclides in a sample is critical for nuclear forensics and emergency response. Radioanalytical techniques are well established for radionuclides measurement, however, they are slow and labor intensive, requiring extensive radiochemical separations and purification prior to analysis. With these limitations of current methods, there is great interest for a new technique to rapidly process samples. This dissertation describes a new analyte extraction medium called Polymer Ligand Film (PLF) developed to rapidly extract radionuclides. Polymer Ligand Film is a polymer medium with ligands incorporated in its matrix that selectively and rapidly extract analytes from a solution. The main focus of the new technique is to shorten and simplify the procedure necessary to chemically isolate radionuclides for determination by alpha spectrometry or beta counting. Five different ligands were tested for plutonium extraction: bis(2-ethylhexyl) methanediphosphonic acid (H2DEH[MDP]), di(2-ethyl hexyl) phosphoric acid (HDEHP), trialkyl methylammonium chloride (Aliquat-336), 4,4'(5')-di-t-butylcyclohexano 18-crown-6 (DtBuCH18C6), and 2-ethylhexyl 2-ethylhexylphosphonic acid (HEH[EHP]). The ligands that were effective for plutonium extraction further studied for uranium extraction. The plutonium recovery by PLFs has shown dependency on nitric acid concentration and ligand to total mass ratio. H2DEH[MDP] PLFs performed best with 1:10 and 1:20 ratio PLFs. 50.44% and 47.61% of plutonium were extracted on the surface of PLFs with 1M nitric acid for 1:10 and 1:20 PLF, respectively. HDEHP PLF provided the best combination of alpha spectroscopy resolution and plutonium recovery with 1:5 PLF when used with 0.1M nitric acid. The overall analyte recovery was lower than electrodeposited samples, which typically has recovery above 80%. However, PLF is designed to be a rapid field deployable screening technique and consistency is more important than recovery. PLFs were also tested using blind quality control samples and the activities were accurately measured. It is important to point out that PLFs were consistently susceptible to analytes penetrating and depositing below the surface. The internal radiation within the body of PLF is mostly contained and did not cause excessive self-attenuation and peak broadening in alpha spectroscopy. The analyte penetration issue was beneficial in the destructive analysis. H2DEH[MDP] PLF was tested with environmental samples to fully understand the capabilities and limitations of the PLF in relevant environments. The extraction system was very effective in extracting plutonium from environmental water collected from Mortandad Canyon at Los Alamos National Laboratory with minimal sample processing. Soil samples were tougher to process than the water samples. Analytes were first leached from the soil matrixes using nitric acid before processing with PLF. This approach had a limitation in extracting plutonium using PLF. The soil samples from Mortandad Canyon, which are about 1% iron by weight, were effectively processed with the PLF system. Even with certain limitations of the PLF extraction system, this technique was able to considerably decrease the sample analysis time. The entire environmental sample was analyzed within one to two days. The decrease in time can be attributed to the fact that PLF is replacing column chromatography and electrodeposition with a single step for preparing alpha spectrometry samples. The two-step process of column chromatography and electrodeposition takes a couple days to a week to complete depending on the sample. The decrease in time and the simplified procedure make this technique a unique solution for application to nuclear forensics and emergency response. A large number of samples can be quickly analyzed and selective samples can be further analyzed with more sensitive techniques based on the initial data. The deployment of a PLF system as a screening method will greatly reduce a total analysis time required to gain meaningful isotopic data for the nuclear forensics application. (Abstract shortened by UMI.)
Klein, Sabrina; Nurjadi, Dennis; Horner, Susanne; Heeg, Klaus; Zimmermann, Stefan; Burckhardt, Irene
2018-04-13
While total laboratory automation (TLA) is well established in laboratory medicine, only a few microbiological laboratories are using TLA systems. Especially in terms of speed and accuracy, working with TLA is expected to be superior to conventional microbiology. We compared in total 35,564 microbiological urine cultures with and without incubation and processing with BD Kiestra TLA for a 6-month period each retrospectively. Sixteen thousand three hundred thirty-eight urine samples were analyzed in the pre-TLA period and 19,226 with TLA. Sixty-two percent (n = 10,101/16338) of the cultures processed without TLA and 68% (n = 13,102/19226) of the cultures processed with TLA showed growth. There were significantly more samples with two or more species per sample and with low numbers of colony forming units (CFU) after incubation with TLA. Regarding the type of bacteria, there were comparable amounts of Enterobacteriaceae in the samples, slightly less non-fermenting Gram-negative bacteria, but significantly more Gram-positive cocci, and Gram-positive rods. Especially Alloscardivia omnicolens, Gardnerella vaginalis, Actinomyces spp., and Actinotignum schaalii were significantly more abundant in the samples incubated and processed with TLA. The time to report was significantly lower in the TLA processed samples by 1.5 h. We provide the first report in Europe of a large number of urine samples processed with TLA. TLA showed enhanced growth of non-classical and rarely cultured bacteria from urine samples. Our findings suggest that previously underestimated bacteria may be relevant pathogens for urinary tract infections. Further studies are needed to confirm our findings.
Optical biosensor system with integrated microfluidic sample preparation and TIRF based detection
NASA Astrophysics Data System (ADS)
Gilli, Eduard; Scheicher, Sylvia R.; Suppan, Michael; Pichler, Heinz; Rumpler, Markus; Satzinger, Valentin; Palfinger, Christian; Reil, Frank; Hajnsek, Martin; Köstler, Stefan
2013-05-01
There is a steadily growing demand for miniaturized bioanalytical devices allowing for on-site or point-of-care detection of biomolecules or pathogens in applications like diagnostics, food testing, or environmental monitoring. These, so called labs-on-a-chip or micro-total analysis systems (μ-TAS) should ideally enable convenient sample-in - result-out type operation. Therefore, the entire process from sample preparation, metering, reagent incubation, etc. to detection should be performed on a single disposable device (on-chip). In the early days such devices were mainly fabricated using glass or silicon substrates and adapting established fabrication technologies from the electronics and semiconductor industry. More recently, the development focuses on the use of thermoplastic polymers as they allow for low-cost high volume fabrication of disposables. One of the most promising materials for the development of plastic based lab-on-achip systems are cyclic olefin polymers and copolymers (COP/COC) due to their excellent optical properties (high transparency and low autofluorescence) and ease of processing. We present a bioanalytical system for whole blood samples comprising a disposable plastic chip based on TIRF (total internal reflection fluorescence) optical detection. The chips were fabricated by compression moulding of COP and microfluidic channels were structured by hot embossing. These microfluidic structures integrate several sample pretreatment steps. These are the separation of erythrocytes, metering of sample volume using passive valves, and reagent incubation for competitive bioassays. The surface of the following optical detection zone is functionalized with specific capture probes in an array format. The plastic chips comprise dedicated structures for simple and effective coupling of excitation light from low-cost laser diodes. This enables TIRF excitation of fluorescently labeled probes selectively bound to detection spots at the microchannel surface. The fluorescence of these detection arrays is imaged using a simple set-up based on a digital consumer camera. Image processing for spot detection and intensity calculation is accomplished using customized software. Using this combined TIRF excitation and imaging based detection approach allowes for effective suppression of background fluorescence from the sample, multiplexed detection in an array format, as well as internal calibration and background correction.
Hinton, Arthur; Cason, J A; Hume, Michael E; Ingram, Kimberly D
2004-08-01
The presence of Campylobacter spp. on broiler carcasses and in scald water taken from a commercial poultry processing facility was monitored on a monthly basis from January through June. Campylobacter agar, Blaser, was used to enumerate Campylobacter in water samples from a multiple-tank scalder; on prescalded, picked, eviscerated, and chilled carcasses; and on processed carcasses stored at 4 degrees C for 7 or 14 days. The MIDI Sherlock microbial identification system was used to identify Campylobacter-like isolates based on the fatty acid methyl ester profile of the bacteria. The dendrogram program of the Sherlock microbial identification system was used to compare the fatty acid methyl ester profiles of the bacteria and determine the degree of relatedness between the isolates. Findings indicated that no Campylobacter were recovered from carcasses or scald tank water samples collected in January or February, but the pathogen was recovered from samples collected in March, April, May, and June. Processing generally produced a significant (P < 0.05) decrease in the number of Campylobacter recovered from broiler carcasses, and the number of Campylobacter recovered from refrigerated carcasses generally decreased during storage. Significantly (P < 0.05) fewer Campylobacter were recovered from the final tank of the multiple-tank scald system than from the first tank. MIDI similarity index values ranged from 0.104 to 0.928 based on MIDI-fatty acid methyl ester analysis of Campylobacterjejuni and Campylobacter coli isolates. Dendrograms of the fatty acid methyl ester profile of the isolates indicated that poultry flocks may introduce several strains of C. jejuni and C. coli into processing plants. Different populations of the pathogen may be carried into the processing plant by successive broiler flocks, and the same Campylobacter strain may be recovered from different poultry processing operations. However, Campylobacter apparently is unable to colonize equipment in the processing facility and contaminate broilers from flocks processed at later dates in the facility.
40 CFR 420.21 - Specialized definitions.
Code of Federal Regulations, 2013 CFR
2013-07-01
... control system that utilizes filters to remove iron-bearing particles (fines) from blast furnace or... level is 10 pg/L per EPA Method 1613B for water and wastewater samples. (d) The term pg/L means... pollution control system means an emission control system that utilizes water to clean process or furnace...
40 CFR 420.21 - Specialized definitions.
Code of Federal Regulations, 2011 CFR
2011-07-01
... control system that utilizes filters to remove iron-bearing particles (fines) from blast furnace or... level is 10 pg/L per EPA Method 1613B for water and wastewater samples. (d) The term pg/L means... pollution control system means an emission control system that utilizes water to clean process or furnace...
40 CFR 420.21 - Specialized definitions.
Code of Federal Regulations, 2014 CFR
2014-07-01
... control system that utilizes filters to remove iron-bearing particles (fines) from blast furnace or... level is 10 pg/L per EPA Method 1613B for water and wastewater samples. (d) The term pg/L means... pollution control system means an emission control system that utilizes water to clean process or furnace...
Pulse-Flow Microencapsulation System
NASA Technical Reports Server (NTRS)
Morrison, Dennis R.
2006-01-01
The pulse-flow microencapsulation system (PFMS) is an automated system that continuously produces a stream of liquid-filled microcapsules for delivery of therapeutic agents to target tissues. Prior microencapsulation systems have relied on batch processes that involve transfer of batches between different apparatuses for different stages of production followed by sampling for acquisition of quality-control data, including measurements of size. In contrast, the PFMS is a single, microprocessor-controlled system that performs all processing steps, including acquisition of quality-control data. The quality-control data can be used as real-time feedback to ensure the production of large quantities of uniform microcapsules.
Modified Pressure System for Imaging Egg Cracks
USDA-ARS?s Scientific Manuscript database
One aspect of grading table eggs is shell checks or cracks. Currently, USDA voluntary regulations require that humans grade a representative sample of all eggs processed. However, as processing plants and packing facilities continue to increase their volume and throughput, human graders are having ...
Modified Pressure System for Imaging Egg Cracks
USDA-ARS?s Scientific Manuscript database
Abstract One aspect of grading table eggs is shell checks or cracks. Currently, USDA voluntary regulations require that humans grade a representative sample of all eggs processed. However, as processing plants and packing facilities continue to increase their volume and throughput, human graders a...
Scale-up of an ultrasound-enhanced bioscouring process
USDA-ARS?s Scientific Manuscript database
Using previously determined optimized reaction conditions, an ultrasound-enhanced bioscouring process was scaled to ten gallon capacity and a system of rollers was added which allowed for continuous fabric feed and equipment operation. UV-Vis photospectroscopic data from bioscoured fabric samples co...
Integrated system for gathering, processing, and reporting data relating to site contamination
Long, D.D.; Goldberg, M.S.; Baker, L.A.
1997-11-11
An integrated screening system comprises an intrusive sampling subsystem, a field mobile laboratory subsystem, a computer assisted design/geographical information subsystem, and a telecommunication linkup subsystem, all integrated to provide synergistically improved data relating to the extent of site soil/groundwater contamination. According to the present invention, data samples related to the soil, groundwater or other contamination of the subsurface material are gathered and analyzed to measure contaminants. Based on the location of origin of the samples in three-dimensional space, the analyzed data are transmitted to a location display. The data from analyzing samples and the data from the locating the origin are managed to project the next probable sample location. The next probable sample location is then forwarded for use as a guide in the placement of ensuing sample location, whereby the number of samples needed to accurately characterize the site is minimized. 10 figs.
Integrated system for gathering, processing, and reporting data relating to site contamination
Long, Delmar D.; Goldberg, Mitchell S.; Baker, Lorie A.
1997-01-01
An integrated screening system comprises an intrusive sampling subsystem, a field mobile laboratory subsystem, a computer assisted design/geographical information subsystem, and a telecommunication linkup subsystem, all integrated to provide synergistically improved data relating to the extent of site soil/groundwater contamination. According to the present invention, data samples related to the soil, groundwater or other contamination of the subsurface material are gathered and analyzed to measure contaminants. Based on the location of origin of the samples in three-dimensional space, the analyzed data are transmitted to a location display. The data from analyzing samples and the data from the locating the origin are managed to project the next probable sample location. The next probable sample location is then forwarded for use as a guide in the placement of ensuing sample location, whereby the number of samples needed to accurately characterize the site is minimized.
Creating ensembles of oblique decision trees with evolutionary algorithms and sampling
Cantu-Paz, Erick [Oakland, CA; Kamath, Chandrika [Tracy, CA
2006-06-13
A decision tree system that is part of a parallel object-oriented pattern recognition system, which in turn is part of an object oriented data mining system. A decision tree process includes the step of reading the data. If necessary, the data is sorted. A potential split of the data is evaluated according to some criterion. An initial split of the data is determined. The final split of the data is determined using evolutionary algorithms and statistical sampling techniques. The data is split. Multiple decision trees are combined in ensembles.
An improved light hydrocarbon analysis system. Interim report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lamontagne, R.A.
1982-05-11
A system for extracting and measuring ambient levels of C1-C4 hydrocarbons and carbon monoxide (CO) in seawater is described. The analytical instrument is a gas chromatograph with flame ionization detectors that incorporates a catalytic conversion of CO to CH4 (methane). The samples are concentrated prior to introduction to the chromatographic system. The volatile hydrocarbons are extracted from the seawater by the use of a helium flow stream and concentrated on dry ice-acetone cold traps. Air samples can be processed in a similar way.
High Accuracy Evaluation of the Finite Fourier Transform Using Sampled Data
NASA Technical Reports Server (NTRS)
Morelli, Eugene A.
1997-01-01
Many system identification and signal processing procedures can be done advantageously in the frequency domain. A required preliminary step for this approach is the transformation of sampled time domain data into the frequency domain. The analytical tool used for this transformation is the finite Fourier transform. Inaccuracy in the transformation can degrade system identification and signal processing results. This work presents a method for evaluating the finite Fourier transform using cubic interpolation of sampled time domain data for high accuracy, and the chirp Zeta-transform for arbitrary frequency resolution. The accuracy of the technique is demonstrated in example cases where the transformation can be evaluated analytically. Arbitrary frequency resolution is shown to be important for capturing details of the data in the frequency domain. The technique is demonstrated using flight test data from a longitudinal maneuver of the F-18 High Alpha Research Vehicle.
Refraction-based X-ray Computed Tomography for Biomedical Purpose Using Dark Field Imaging Method
NASA Astrophysics Data System (ADS)
Sunaguchi, Naoki; Yuasa, Tetsuya; Huo, Qingkai; Ichihara, Shu; Ando, Masami
We have proposed a tomographic x-ray imaging system using DFI (dark field imaging) optics along with a data-processing method to extract information on refraction from the measured intensities, and a reconstruction algorithm to reconstruct a refractive-index field from the projections generated from the extracted refraction information. The DFI imaging system consists of a tandem optical system of Bragg- and Laue-case crystals, a positioning device system for a sample, and two CCD (charge coupled device) cameras. Then, we developed a software code to simulate the data-acquisition, data-processing, and reconstruction methods to investigate the feasibility of the proposed methods. Finally, in order to demonstrate its efficacy, we imaged a sample with DCIS (ductal carcinoma in situ) excised from a breast cancer patient using a system constructed at the vertical wiggler beamline BL-14C in KEK-PF. Its CT images depicted a variety of fine histological structures, such as milk ducts, duct walls, secretions, adipose and fibrous tissue. They correlate well with histological sections.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Linard, Joshua; Campbell, Sam
This event included annual sampling of groundwater and surface water locations at the Gunnison, Colorado, Processing Site. Sampling and analyses were conducted as specified in Sampling and Analysis Plan for US Department of Energy Office of Legacy Management Sites (LMS/PRO/S04351, continually updated, http://energy.gov/lm/downloads/sampling-and analysis-plan-us-department-energy-office-legacy-management-sites). Samples were collected from 28 monitoring wells, three domestic wells, and six surface locations in April at the processing site as specified in the draft 2010 Ground Water Compliance Action Plan for the Gunnison, Colorado, Processing Site. Planned monitoring locations are shown in Attachment 1, Sampling and Analysis Work Order. Domestic wells 0476 and 0477 weremore » sampled in June because the homes were unoccupied in April, and the wells were not in use. Duplicate samples were collected from locations 0126, 0477, and 0780. One equipment blank was collected during this sampling event. Water levels were measured at all monitoring wells that were sampled. See Attachment 2, Trip Reports for additional details. The analytical data and associated qualifiers can be viewed in environmental database reports and are also available for viewing with dynamic mapping via the GEMS (Geospatial Environmental Mapping System) website at http://gems.lm.doe.gov/#. No issues were identified during the data validation process that requires additional action or follow-up. An assessment of anomalous data is included in Attachment 3. Interpretation and presentation of results, including an assessment ofthe natural flushing compliance strategy, will be reported in the upcoming 2016 Verification Monitoring Report. U.S.« less
Kim, Jungkyu; Jensen, Erik C; Stockton, Amanda M; Mathies, Richard A
2013-08-20
A fully integrated multilayer microfluidic chemical analyzer for automated sample processing and labeling, as well as analysis using capillary zone electrophoresis is developed and characterized. Using lifting gate microfluidic control valve technology, a microfluidic automaton consisting of a two-dimensional microvalve cellular array is fabricated with soft lithography in a format that enables facile integration with a microfluidic capillary electrophoresis device. The programmable sample processor performs precise mixing, metering, and routing operations that can be combined to achieve automation of complex and diverse assay protocols. Sample labeling protocols for amino acid, aldehyde/ketone and carboxylic acid analysis are performed automatically followed by automated transfer and analysis by the integrated microfluidic capillary electrophoresis chip. Equivalent performance to off-chip sample processing is demonstrated for each compound class; the automated analysis resulted in a limit of detection of ~16 nM for amino acids. Our microfluidic automaton provides a fully automated, portable microfluidic analysis system capable of autonomous analysis of diverse compound classes in challenging environments.
Physical Processing of Cometary Nuclei
NASA Technical Reports Server (NTRS)
Weissman, Paul R.; Stern, S. Alan
1997-01-01
Cometary nuclei preserve a cosmo-chemical record of conditions and processes in the primordial solar nebula, and possibly even the interstellar medium. However, that record is not perfectly preserved over the age of the solar system due to a variety of physical processes which act to modify cometary surfaces and interiors. Possible structural and/or internal processes include: collisional accretion, disruption, and reassembly during formation; internal heating by long and short-lived radionuclides; amorphous to crystalline phase transitions, and thermal stresses. Identified surface modification processes include: irradiation by galactic cosmic rays, solar protons, UV photons, and the Sun's T Tauri stage mass outflow; heating by passing stars and nearby supernovae; gardening by debris impacts; the accretion of interstellar dust and gas and accompanying erosion by hypervelocity dust impacts and sputtering; and solar heating with accompanying crust formation. These modification processes must be taken into account in both the planning and the interpretation of the results of a Comet Nucleus Sample Return Mission. Sampling of nuclei should be done at as great a depth below the surface crust as technically feasible, and at vents or fissures leading to exposed volatiles at depth. Samples of the expected cometary crust and near-surface layers also need to be returned for analysis to achieve a better understanding of the effects of these physical processes. We stress that comets are still likely less modified dm any other solar system bodies, but the degree of modification can vary greatly from one comet to the next.
Microfluidic Sample Preparation for Diagnostic Cytopathology
Mach, Albert J.; Adeyiga, Oladunni B.; Di Carlo, Dino
2014-01-01
The cellular components of body fluids are routinely analyzed to identify disease and treatment approaches. While significant focus has been placed on developing cell analysis technologies, tools to automate the preparation of cellular specimens have been more limited, especially for body fluids beyond blood. Preparation steps include separating, concentrating, and exposing cells to reagents. Sample preparation continues to be routinely performed off-chip by technicians, preventing cell-based point-of-care diagnostics, increasing the cost of tests, and reducing the consistency of the final analysis following multiple manually-performed steps. Here, we review the assortment of biofluids for which suspended cells are analyzed, along with their characteristics and diagnostic value. We present an overview of the conventional sample preparation processes for cytological diagnosis. We finally discuss the challenges and opportunities in developing microfluidic devices for the purpose of automating or miniaturizing these processes, with particular emphases on preparing large or small volume samples, working with samples of high cellularity, automating multi-step processes, and obtaining high purity subpopulations of cells. We hope to convey the importance of and help identify new research directions addressing the vast biological and clinical applications in preparing and analyzing the array of available biological fluids. Successfully addressing the challenges described in this review can lead to inexpensive systems to improve diagnostic accuracy while simultaneously reducing overall systemic healthcare costs. PMID:23380972
Research on width control of Metal Fused-coating Additive Manufacturing based on active control
NASA Astrophysics Data System (ADS)
Ren, Chuan qi; Wei, Zheng ying; Wang, Xin; Du, Jun; Zhang, Shan; Zhang, Zhitong; Bai, Hao
2017-12-01
Given the stability of the shape of the forming layer is one of the key problems that affect the final quality of the sample morphology, taking a study on the forming process and the control method of morphology make a significant difference to metal fused-coating additive manufacturing (MFCAM) in achieving the efficient and stable forming. To improve the quality and precision of the samples of single-layer single pass, a control method of morphology based on active control was established by this paper. The real-time acquisition of image was realized by CCD and the characteristics of morphology of the forming process were simultaneously extracted. Making analysis of the characteristics of the width during the process, the relationship between the relative difference of different frames and moving speed was given. A large number of experiments are used to verify the response speed and accuracy of the system. The results show that the active system can improve the morphology of the sample and the smoothness of the width of the single channel, and increase the uniformity of width by 55.16%.
The OSIRIS-REx Mission Sample Site Selection Process
NASA Astrophysics Data System (ADS)
Beshore, Edward C.; Lauretta, Dante
2014-11-01
In September of 2016, the OSIRIS-REx (Origins, Spectral Interpretation, Resource Identification, Security, REgolith eXplorer) spacecraft will depart for asteroid (101955) Bennu, and in doing so, will turn an important corner in the exploration of the solar system. After arriving at Bennu in the fall of 2018, OSIRIS-REx will undertake a program of observations designed to select a site suitable for retrieving a sample that will be returned to the Earth in 2023. The third mission in NASA’s New Frontiers program, OSIRIS-REx will return over 60 grams from Bennu’s surface.OSIRIS-REx is unique because the science team will have an operational role to play in preparing data products needed to select a sample site. These include products used to ensure flight system safety — topographic maps and shape models, temperature measurements, maps of hazards — as well as assessments of sampleability and science value. The timing and production of these will be presented, as will the high-level decision-making tools and processes for the interim and final site selection processes.
McCloy, J S; Sundaram, S K; Matyas, J; Woskov, P P
2011-05-01
Millimeter wave (MMW) radiometry can be used for simultaneous measurement of emissivity and temperature of materials under extreme environments (high temperature, pressure, and corrosive environments). The state-of-the-art dual channel MMW passive radiometer with active interferometric capabilities at 137 GHz described here allows for radiometric measurements of sample temperature and emissivity up to at least 1600 °C with simultaneous measurement of sample surface dynamics. These capabilities have been used to demonstrate dynamic measurement of melting of powders of simulated lunar regolith and static measurement of emissivity of solid samples. The paper presents the theoretical background and basis for the dual-receiver system, describes the hardware in detail, and demonstrates the data analysis. Post-experiment analysis of emissivity versus temperature allows further extraction from the radiometric data of millimeter wave viewing beam coupling factors, which provide corroboratory evidence to the interferometric data of the process dynamics observed. These results show the promise of the MMW system for extracting quantitative and qualitative process parameters for industrial processes and access to real-time dynamics of materials behavior in extreme environments.
Development of a fast framing detector for electron microscopy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Johnson, Ian J.; Bustillo, Karen C.; Ciston, Jim
2016-10-01
A high frame rate detector system is described that enables fast real-time data analysis of scanning diffraction experiments in scanning transmission electron microscopy (STEM). This is an end-to-end development that encompasses the data producing detector, data transportation, and real-time processing of data. The detector will consist of a central pixel sensor that is surrounded by annular silicon diodes. Both components of the detector system will synchronously capture data at almost 100 kHz frame rate, which produces an approximately 400 Gb/s data stream. Low-level preprocessing will be implemented in firmware before the data is streamed from the National Center for Electronmore » Microscopy (NCEM) to the National Energy Research Scientific Computing Center (NERSC). Live data processing, before it lands on disk, will happen on the Cori supercomputer and aims to present scientists with prompt experimental feedback. This online analysis will provide rough information of the sample that can be utilized for sample alignment, sample monitoring and verification that the experiment is set up correctly. Only a compressed version of the relevant data is then selected for more in-depth processing.« less
Smart Networked Elements in Support of ISHM
NASA Technical Reports Server (NTRS)
Oostdyk, Rebecca; Mata, Carlos; Perotti, Jose M.
2008-01-01
At the core of ISHM is the ability to extract information and knowledge from raw data. Conventional data acquisition systems sample and convert physical measurements to engineering units, which higher-level systems use to derive health and information about processes and systems. Although health management is essential at the top level, there are considerable advantages to implementing health-related functions at the sensor level. The distribution of processing to lower levels reduces bandwidth requirements, enhances data fusion, and improves the resolution for detection and isolation of failures in a system, subsystem, component, or process. The Smart Networked Element (SNE) has been developed to implement intelligent functions and algorithms at the sensor level in support of ISHM.
A noninvasive technique for real-time detection of bruises in apple surface based on machine vision
NASA Astrophysics Data System (ADS)
Zhao, Juan; Peng, Yankun; Dhakal, Sagar; Zhang, Leilei; Sasao, Akira
2013-05-01
Apple is one of the highly consumed fruit item in daily life. However, due to its high damage potential and massive influence on taste and export, the quality of apple has to be detected before it reaches the consumer's hand. This study was aimed to develop a hardware and software unit for real-time detection of apple bruises based on machine vision technology. The hardware unit consisted of a light shield installed two monochrome cameras at different angles, LED light source to illuminate the sample, and sensors at the entrance of box to signal the positioning of sample. Graphical Users Interface (GUI) was developed in VS2010 platform to control the overall hardware and display the image processing result. The hardware-software system was developed to acquire the images of 3 samples from each camera and display the image processing result in real time basis. An image processing algorithm was developed in Opencv and C++ platform. The software is able to control the hardware system to classify the apple into two grades based on presence/absence of surface bruises with the size of 5mm. The experimental result is promising and the system with further modification can be applicable for industrial production in near future.
Salceda, Susana; Barican, Arnaldo; Buscaino, Jacklyn; Goldman, Bruce; Klevenberg, Jim; Kuhn, Melissa; Lehto, Dennis; Lin, Frank; Nguyen, Phong; Park, Charles; Pearson, Francesca; Pittaro, Rick; Salodkar, Sayali; Schueren, Robert; Smith, Corey; Troup, Charles; Tsou, Dean; Vangbo, Mattias; Wunderle, Justus; King, David
2017-05-01
The RapidHIT ® ID is a fully automated sample-to-answer system for short tandem repeat (STR)-based human identification. The RapidHIT ID has been optimized for use in decentralized environments and processes presumed single source DNA samples, generating Combined DNA Index System (CODIS)-compatible DNA profiles in less than 90min. The system is easy to use, requiring less than one minute of hands-on time. Profiles are reviewed using centralized linking software, RapidLINK™ (IntegenX, Pleasanton, CA), a software tool designed to collate DNA profiles from single or multiple RapidHIT ID systems at different geographic locations. The RapidHIT ID has been designed to employ GlobalFiler ® Express and AmpFLSTR ® NGMSElect™, Thermo Fisher Scientific (Waltham, MA) STR chemistries. The Developmental Validation studies were performed using GlobalFiler ® Express with single source reference samples according to Scientific Working Group for DNA Analysis Methods guidelines. These results show that multiple RapidHIT ID systems networked with RapidLINK software form a highly reliable system for wide-scale deployment in locations such as police booking stations and border crossings enabling real-time testing of arrestees, potential human trafficking victims, and other instances where rapid turnaround is essential. Copyright © 2017 The Author(s). Published by Elsevier B.V. All rights reserved.
Synthesis of carbon nanofibers by catalytic CVD of chlorobenzene over bulk nickel alloy
NASA Astrophysics Data System (ADS)
Kenzhin, Roman M.; Bauman, Yuri I.; Volodin, Alexander M.; Mishakov, Ilya V.; Vedyagin, Aleksey A.
2018-01-01
Catalytic chemical vapor deposition (CCVD) of chlorobenzene over bulk nickel alloy (nichrome) was studied. The bulk Ni-containing samples being exposed to a contact with aggressive reaction medium undergo self-disintegration followed by growth of carbon nanofibers. This process, also known as a metal dusting, requires the simultaneous presence of chlorine and hydrogen sources in the reaction mixture. Molecule of chlorobenzene complies with these requirements. The experiments on CCVD were performed in a flow-through reactor system. The initial stages of nickel disintegration process were investigated in a closed system under Autogenic Pressure at Elevated Temperature (RAPET) conditions. Scanning and transmission electron microscopies and ferromagnetic resonance spectroscopy were applied to examine the samples after their interaction with chlorobenzene. Introduction of additional hydrogen into the flow-through system was shown to affect the morphology of grown carbon nanofibers.
Adaptive model predictive process control using neural networks
Buescher, K.L.; Baum, C.C.; Jones, R.D.
1997-08-19
A control system for controlling the output of at least one plant process output parameter is implemented by adaptive model predictive control using a neural network. An improved method and apparatus provides for sampling plant output and control input at a first sampling rate to provide control inputs at the fast rate. The MPC system is, however, provided with a network state vector that is constructed at a second, slower rate so that the input control values used by the MPC system are averaged over a gapped time period. Another improvement is a provision for on-line training that may include difference training, curvature training, and basis center adjustment to maintain the weights and basis centers of the neural in an updated state that can follow changes in the plant operation apart from initial off-line training data. 46 figs.
Adaptive model predictive process control using neural networks
Buescher, Kevin L.; Baum, Christopher C.; Jones, Roger D.
1997-01-01
A control system for controlling the output of at least one plant process output parameter is implemented by adaptive model predictive control using a neural network. An improved method and apparatus provides for sampling plant output and control input at a first sampling rate to provide control inputs at the fast rate. The MPC system is, however, provided with a network state vector that is constructed at a second, slower rate so that the input control values used by the MPC system are averaged over a gapped time period. Another improvement is a provision for on-line training that may include difference training, curvature training, and basis center adjustment to maintain the weights and basis centers of the neural in an updated state that can follow changes in the plant operation apart from initial off-line training data.
NASA Astrophysics Data System (ADS)
Peters, Stefan T. M.; Münker, Carsten; Pfeifer, Markus; Elfers, Bo-Magnus; Sprung, Peter
2017-02-01
Some nuclides that were produced in supernovae are heterogeneously distributed between different meteoritic materials. In some cases these heterogeneities have been interpreted as the result of interaction between ejecta from a nearby supernova and the nascent solar system. Particularly in the case of the oldest objects that formed in the solar system - Ca-Al rich inclusions (CAIs) - this view is confirm the hypothesis that a nearby supernova event facilitated or even triggered solar system formation. We present Hf isotope data for bulk meteorites, terrestrial materials and CAIs, for the first time including the low-abundance isotope 174Hf (∼0.16%). This rare isotope was likely produced during explosive O/Ne shell burning in massive stars (i.e., the classical "p-process"), and therefore its abundance potentially provides a sensitive tracer for putative heterogeneities within the solar system that were introduced by supernova ejecta. For CAIs and one LL chondrite, also complementary W isotope data are reported for the same sample cuts. Once corrected for small neutron capture effects, different chondrite groups, eucrites, a silicate inclusion of a IAB iron meteorite, and terrestrial materials display homogeneous Hf isotope compositions including 174Hf. Hafnium-174 was thus uniformly distributed in the inner solar system when planetesimals formed at the <50 ppm level. This finding is in good agreement with the evidently homogeneous distributions of p-process isotopes 180W, 184Os and possibly 190Pt between different iron meteorite groups. In contrast to bulk meteorite samples, CAIs show variable depletions in p-process 174Hf with respect to the inner solar system composition, and also variable r-process (or s-process) Hf and W contributions. Based on combined Hf and W isotope compositions, we show that CAIs sampled at least one component in which the proportion of r- and s-process derived Hf and W deviates from that of supernova ejecta. The Hf and W isotope anomalies in CAIs are therefore best explained by selective processing of presolar carrier phases prior to CAI formation, and not by a late injection of supernova materials. Likewise, other isotope anomalies in additional elements in CAIs relative to the bulk solar system may reflect the same process. The isotopic heterogeneities between the first refractory condensates may have been eradicated partially during CAI formation, because W isotope anomalies in CAIs appear to decrease with increasing W concentrations as inferred from time-integrated 182W/184W. Importantly, the 176Lu-176Hf and 182Hf-182W chronometers are not significantly affected by nucleosynthetic heterogeneity of Hf isotopes in bulk meteorites, but may be affected in CAIs.
Calabria, Andrea; Spinozzi, Giulio; Benedicenti, Fabrizio; Tenderini, Erika; Montini, Eugenio
2015-01-01
Many biological laboratories that deal with genomic samples are facing the problem of sample tracking, both for pure laboratory management and for efficiency. Our laboratory exploits PCR techniques and Next Generation Sequencing (NGS) methods to perform high-throughput integration site monitoring in different clinical trials and scientific projects. Because of the huge amount of samples that we process every year, which result in hundreds of millions of sequencing reads, we need to standardize data management and tracking systems, building up a scalable and flexible structure with web-based interfaces, which are usually called Laboratory Information Management System (LIMS). We started collecting end-users' requirements, composed of desired functionalities of the system and Graphical User Interfaces (GUI), and then we evaluated available tools that could address our requirements, spanning from pure LIMS to Content Management Systems (CMS) up to enterprise information systems. Our analysis identified ADempiere ERP, an open source Enterprise Resource Planning written in Java J2EE, as the best software that also natively implements some highly desirable technological advances, such as the high usability and modularity that grants high use-case flexibility and software scalability for custom solutions. We extended and customized ADempiere ERP to fulfil LIMS requirements and we developed adLIMS. It has been validated by our end-users verifying functionalities and GUIs through test cases for PCRs samples and pre-sequencing data and it is currently in use in our laboratories. adLIMS implements authorization and authentication policies, allowing multiple users management and roles definition that enables specific permissions, operations and data views to each user. For example, adLIMS allows creating sample sheets from stored data using available exporting operations. This simplicity and process standardization may avoid manual errors and information backtracking, features that are not granted using track recording on files or spreadsheets. adLIMS aims to combine sample tracking and data reporting features with higher accessibility and usability of GUIs, thus allowing time to be saved on doing repetitive laboratory tasks, and reducing errors with respect to manual data collection methods. Moreover, adLIMS implements automated data entry, exploiting sample data multiplexing and parallel/transactional processing. adLIMS is natively extensible to cope with laboratory automation through platform-dependent API interfaces, and could be extended to genomic facilities due to the ERP functionalities.
Shibata, Kazuhiro; Itoh, Masayoshi; Aizawa, Katsunori; Nagaoka, Sumiharu; Sasaki, Nobuya; Carninci, Piero; Konno, Hideaki; Akiyama, Junichi; Nishi, Katsuo; Kitsunai, Tokuji; Tashiro, Hideo; Itoh, Mari; Sumi, Noriko; Ishii, Yoshiyuki; Nakamura, Shin; Hazama, Makoto; Nishine, Tsutomu; Harada, Akira; Yamamoto, Rintaro; Matsumoto, Hiroyuki; Sakaguchi, Sumito; Ikegami, Takashi; Kashiwagi, Katsuya; Fujiwake, Syuji; Inoue, Kouji; Togawa, Yoshiyuki; Izawa, Masaki; Ohara, Eiji; Watahiki, Masanori; Yoneda, Yuko; Ishikawa, Tomokazu; Ozawa, Kaori; Tanaka, Takumi; Matsuura, Shuji; Kawai, Jun; Okazaki, Yasushi; Muramatsu, Masami; Inoue, Yorinao; Kira, Akira; Hayashizaki, Yoshihide
2000-01-01
The RIKEN high-throughput 384-format sequencing pipeline (RISA system) including a 384-multicapillary sequencer (the so-called RISA sequencer) was developed for the RIKEN mouse encyclopedia project. The RISA system consists of colony picking, template preparation, sequencing reaction, and the sequencing process. A novel high-throughput 384-format capillary sequencer system (RISA sequencer system) was developed for the sequencing process. This system consists of a 384-multicapillary auto sequencer (RISA sequencer), a 384-multicapillary array assembler (CAS), and a 384-multicapillary casting device. The RISA sequencer can simultaneously analyze 384 independent sequencing products. The optical system is a scanning system chosen after careful comparison with an image detection system for the simultaneous detection of the 384-capillary array. This scanning system can be used with any fluorescent-labeled sequencing reaction (chain termination reaction), including transcriptional sequencing based on RNA polymerase, which was originally developed by us, and cycle sequencing based on thermostable DNA polymerase. For long-read sequencing, 380 out of 384 sequences (99.2%) were successfully analyzed and the average read length, with more than 99% accuracy, was 654.4 bp. A single RISA sequencer can analyze 216 kb with >99% accuracy in 2.7 h (90 kb/h). For short-read sequencing to cluster the 3′ end and 5′ end sequencing by reading 350 bp, 384 samples can be analyzed in 1.5 h. We have also developed a RISA inoculator, RISA filtrator and densitometer, RISA plasmid preparator which can handle throughput of 40,000 samples in 17.5 h, and a high-throughput RISA thermal cycler which has four 384-well sites. The combination of these technologies allowed us to construct the RISA system consisting of 16 RISA sequencers, which can process 50,000 DNA samples per day. One haploid genome shotgun sequence of a higher organism, such as human, mouse, rat, domestic animals, and plants, can be revealed by seven RISA systems within one month. PMID:11076861
Modification of a Microwave Oven for Laboratory Use.
ERIC Educational Resources Information Center
Andrews, Judith; Atkinson, George F.
1984-01-01
Discusses use of a domestic microwave oven for drying analytical samples with time savings compared to conventional ovens, providing a solution to the problem of loss of load as samples dry. Presents a system for examining emitted gases from drying process and reports results of several test dryings. (JM)
Microfluidics-to-Mass Spectrometry: A review of coupling methods and applications
Wang, Xue; Yi, Lian; Mukhitov, Nikita; Schrell, Adrian M.; Dhumpa, Raghuram; Roper, Michael G.
2014-01-01
Microfluidic devices offer great advantages in integrating sample processes, minimizing sample and reagent volumes, and increasing analysis speed, while mass spectrometry detection provides high information content, is sensitive, and can be used in quantitative analyses. The coupling of microfluidic devices to mass spectrometers is becoming more common with the strengths of both systems being combined to analyze precious and complex samples. This review summarizes select achievements published between 2010 – July 2014 in novel coupling between microfluidic devices and mass spectrometers. The review is subdivided by the types of ionization sources employed, and the different microfluidic systems used. PMID:25458901
Sampling and data handling methods for inhalable particulate sampling. Final report nov 78-dec 80
DOE Office of Scientific and Technical Information (OSTI.GOV)
Smith, W.B.; Cushing, K.M.; Johnson, J.W.
1982-05-01
The report reviews the objectives of a research program on sampling and measuring particles in the inhalable particulate (IP) size range in emissions from stationary sources, and describes methods and equipment required. A computer technique was developed to analyze data on particle-size distributions of samples taken with cascade impactors from industrial process streams. Research in sampling systems for IP matter included concepts for maintaining isokinetic sampling conditions, necessary for representative sampling of the larger particles, while flowrates in the particle-sizing device were constant. Laboratory studies were conducted to develop suitable IP sampling systems with overall cut diameters of 15 micrometersmore » and conforming to a specified collection efficiency curve. Collection efficiencies were similarly measured for a horizontal elutriator. Design parameters were calculated for horizontal elutriators to be used with impactors, the EPA SASS train, and the EPA FAS train. Two cyclone systems were designed and evaluated. Tests on an Andersen Size Selective Inlet, a 15-micrometer precollector for high-volume samplers, showed its performance to be with the proposed limits for IP samplers. A stack sampling system was designed in which the aerosol is diluted in flow patterns and with mixing times simulating those in stack plumes.« less
Perskvist, Nasrin; Norlin, Loreana; Dillner, Joakim
2015-04-01
This article addresses the important issue of the standardization of the biobank process. It reports on i) the implementation of standard operating procedures for the processing of liquid-based cervical cells, ii) the standardization of storage conditions, and iii) the ultimate establishment of nationwide standardized biorepositories for cervical specimens. Given the differences in the infrastructure and healthcare systems of various county councils in Sweden, these efforts were designed to develop standardized methods of biobanking across the nation. The standardization of cervical sample processing and biobanking is an important and widely acknowledged issue. Efforts to address these concerns will facilitate better patient care and improve research based on retrospective and prospective collections of patient samples and cohorts. The successful nationalization of the Cervical Cytology Biobank in Sweden is based on three vital issues: i) the flexibility of the system to adapt to other regional systems, ii) the development of the system based on national collaboration between the university and the county councils, and iii) stable governmental financing by the provider, the Biobanking and Molecular Resource Infrastructure of Sweden (BBMRI.se). We will share our experiences with biorepository communities to promote understanding of and advances in opportunities to establish a nationalized biobank which covers the healthcare of the entire nation.
Pursley, Randall H.; Salem, Ghadi; Devasahayam, Nallathamby; Subramanian, Sankaran; Koscielniak, Janusz; Krishna, Murali C.; Pohida, Thomas J.
2006-01-01
The integration of modern data acquisition and digital signal processing (DSP) technologies with Fourier transform electron paramagnetic resonance (FT-EPR) imaging at radiofrequencies (RF) is described. The FT-EPR system operates at a Larmor frequency (Lf) of 300 MHz to facilitate in vivo studies. This relatively low frequency Lf, in conjunction with our ~10 MHz signal bandwidth, enables the use of direct free induction decay time-locked subsampling (TLSS). This particular technique provides advantages by eliminating the traditional analog intermediate frequency downconversion stage along with the corresponding noise sources. TLSS also results in manageable sample rates that facilitate the design of DSP-based data acquisition and image processing platforms. More specifically, we utilize a high-speed field programmable gate array (FPGA) and a DSP processor to perform advanced real-time signal and image processing. The migration to a DSP-based configuration offers the benefits of improved EPR system performance, as well as increased adaptability to various EPR system configurations (i.e., software configurable systems instead of hardware reconfigurations). The required modifications to the FT-EPR system design are described, with focus on the addition of DSP technologies including the application-specific hardware, software, and firmware developed for the FPGA and DSP processor. The first results of using real-time DSP technologies in conjunction with direct detection bandpass sampling to implement EPR imaging at RF frequencies are presented. PMID:16243552
Two phase sampling for wheat acreage estimation. [large area crop inventory experiment
NASA Technical Reports Server (NTRS)
Thomas, R. W.; Hay, C. M.
1977-01-01
A two phase LANDSAT-based sample allocation and wheat proportion estimation method was developed. This technique employs manual, LANDSAT full frame-based wheat or cultivated land proportion estimates from a large number of segments comprising a first sample phase to optimally allocate a smaller phase two sample of computer or manually processed segments. Application to the Kansas Southwest CRD for 1974 produced a wheat acreage estimate for that CRD within 2.42 percent of the USDA SRS-based estimate using a lower CRD inventory budget than for a simulated reference LACIE system. Factor of 2 or greater cost or precision improvements relative to the reference system were obtained.
Underwater microscope for measuring spatial and temporal changes in bed-sediment grain size
Rubin, David M.; Chezar, Henry; Harney, Jodi N.; Topping, David J.; Melis, Theodore S.; Sherwood, Christopher R.
2007-01-01
For more than a century, studies of sedimentology and sediment transport have measured bed-sediment grain size by collecting samples and transporting them back to the laboratory for grain-size analysis. This process is slow and expensive. Moreover, most sampling systems are not selective enough to sample only the surficial grains that interact with the flow; samples typically include sediment from at least a few centimeters beneath the bed surface. New hardware and software are available for in situ measurement of grain size. The new technology permits rapid measurement of surficial bed sediment. Here we describe several systems we have deployed by boat, by hand, and by tripod in rivers, oceans, and on beaches.
Underwater Microscope for Measuring Spatial and Temporal Changes in Bed-Sediment Grain Size
Rubin, David M.; Chezar, Henry; Harney, Jodi N.; Topping, David J.; Melis, Theodore S.; Sherwood, Christopher R.
2006-01-01
For more than a century, studies of sedimentology and sediment transport have measured bed-sediment grain size by collecting samples and transporting them back to the lab for grain-size analysis. This process is slow and expensive. Moreover, most sampling systems are not selective enough to sample only the surficial grains that interact with the flow; samples typically include sediment from at least a few centimeters beneath the bed surface. New hardware and software are available for in-situ measurement of grain size. The new technology permits rapid measurement of surficial bed sediment. Here we describe several systems we have deployed by boat, by hand, and by tripod in rivers, oceans, and on beaches.
Gavrielides, Mike; Furney, Simon J; Yates, Tim; Miller, Crispin J; Marais, Richard
2014-01-01
Whole genomes, whole exomes and transcriptomes of tumour samples are sequenced routinely to identify the drivers of cancer. The systematic sequencing and analysis of tumour samples, as well other oncogenomic experiments, necessitates the tracking of relevant sample information throughout the investigative process. These meta-data of the sequencing and analysis procedures include information about the samples and projects as well as the sequencing centres, platforms, data locations, results locations, alignments, analysis specifications and further information relevant to the experiments. The current work presents a sample tracking system for oncogenomic studies (Onco-STS) to store these data and make them easily accessible to the researchers who work with the samples. The system is a web application, which includes a database and a front-end web page that allows the remote access, submission and updating of the sample data in the database. The web application development programming framework Grails was used for the development and implementation of the system. The resulting Onco-STS solution is efficient, secure and easy to use and is intended to replace the manual data handling of text records. Onco-STS allows simultaneous remote access to the system making collaboration among researchers more effective. The system stores both information on the samples in oncogenomic studies and details of the analyses conducted on the resulting data. Onco-STS is based on open-source software, is easy to develop and can be modified according to a research group's needs. Hence it is suitable for laboratories that do not require a commercial system.
NASA Astrophysics Data System (ADS)
Meyer, F. J.; McAlpin, D. B.; Gong, W.; Ajadi, O.; Arko, S.; Webley, P. W.; Dehn, J.
2015-02-01
Remote sensing plays a critical role in operational volcano monitoring due to the often remote locations of volcanic systems and the large spatial extent of potential eruption pre-cursor signals. Despite the all-weather capabilities of radar remote sensing and its high performance in monitoring of change, the contribution of radar data to operational monitoring activities has been limited in the past. This is largely due to: (1) the high costs associated with radar data; (2) traditionally slow data processing and delivery procedures; and (3) the limited temporal sampling provided by spaceborne radars. With this paper, we present new data processing and data integration techniques that mitigate some of these limitations and allow for a meaningful integration of radar data into operational volcano monitoring decision support systems. Specifically, we present fast data access procedures as well as new approaches to multi-track processing that improve near real-time data access and temporal sampling of volcanic systems with SAR data. We introduce phase-based (coherent) and amplitude-based (incoherent) change detection procedures that are able to extract dense time series of hazard information from these data. For a demonstration, we present an integration of our processing system with an operational volcano monitoring system that was developed for use by the Alaska Volcano Observatory (AVO). Through an application to a historic eruption, we show that the integration of SAR into systems such as AVO can significantly improve the ability of operational systems to detect eruptive precursors. Therefore, the developed technology is expected to improve operational hazard detection, alerting, and management capabilities.
Aging near the wall in colloidal glasses
NASA Astrophysics Data System (ADS)
Cao, Cong; Huang, Xinru; Weeks, Eric
In a colloidal glass system, particles move slower as sample ages. In addition, their motions may be affected by their local structure, and this structure will be different near a wall. We examine how the aging process near a wall differs from that in the bulk of the sample. In particular, we use a confocal microscope to observe 3D motion in a bidisperse colloidal glass sample. We find that flat walls induce the particles to organize into layers. The aging process behaves differently near the boundary, especially within the first three layers. Particle motion near the wall is noticeably slower but also changes less dramatically with age. We compare and contrast aging seen in samples with flat and rough walls.
Application of automation and information systems to forensic genetic specimen processing.
Leclair, Benoît; Scholl, Tom
2005-03-01
During the last 10 years, the introduction of PCR-based DNA typing technologies in forensic applications has been highly successful. This technology has become pervasive throughout forensic laboratories and it continues to grow in prevalence. For many criminal cases, it provides the most probative evidence. Criminal genotype data banking and victim identification initiatives that follow mass-fatality incidents have benefited the most from the introduction of automation for sample processing and data analysis. Attributes of offender specimens including large numbers, high quality and identical collection and processing are ideal for the application of laboratory automation. The magnitude of kinship analysis required by mass-fatality incidents necessitates the application of computing solutions to automate the task. More recently, the development activities of many forensic laboratories are focused on leveraging experience from these two applications to casework sample processing. The trend toward increased prevalence of forensic genetic analysis will continue to drive additional innovations in high-throughput laboratory automation and information systems.
Effect of Cell-seeded Hydroxyapatite Scaffolds on Rabbit Radius Bone Regeneration
2013-06-22
OK) for 14 d via a tissue processer (Leica TP1020 system; Bannockburn, IL). Samples were then embedded in photocuring resin (Technovit 7200 VLC ...Kulzer, Germany) and polymerized under blue light for 24 h. Block samples were adhered to a parallel plexiglass slide using the Exakt 7210 VLC system...induction, choice of evaluation time point, and use of a nonhealing defect. For example, a more challenging radial defect (1.5 cm) in rabbits and the
Internet-based profiler system as integrative framework to support translational research
Kim, Robert; Demichelis, Francesca; Tang, Jeffery; Riva, Alberto; Shen, Ronglai; Gibbs, Doug F; Mahavishno, Vasudeva; Chinnaiyan, Arul M; Rubin, Mark A
2005-01-01
Background Translational research requires taking basic science observations and developing them into clinically useful tests and therapeutics. We have developed a process to develop molecular biomarkers for diagnosis and prognosis by integrating tissue microarray (TMA) technology and an internet-database tool, Profiler. TMA technology allows investigators to study hundreds of patient samples on a single glass slide resulting in the conservation of tissue and the reduction in inter-experimental variability. The Profiler system allows investigator to reliably track, store, and evaluate TMA experiments. Here within we describe the process that has evolved through an empirical basis over the past 5 years at two academic institutions. Results The generic design of this system makes it compatible with multiple organ system (e.g., prostate, breast, lung, renal, and hematopoietic system,). Studies and folders are restricted to authorized users as required. Over the past 5 years, investigators at 2 academic institutions have scanned 656 TMA experiments and collected 63,311 digital images of these tissue samples. 68 pathologists from 12 major user groups have accessed the system. Two groups directly link clinical data from over 500 patients for immediate access and the remaining groups choose to maintain clinical and pathology data on separate systems. Profiler currently has 170 K data points such as staining intensity, tumor grade, and nuclear size. Due to the relational database structure, analysis can be easily performed on single or multiple TMA experimental results. The TMA module of Profiler can maintain images acquired from multiple systems. Conclusion We have developed a robust process to develop molecular biomarkers using TMA technology and an internet-based database system to track all steps of this process. This system is extendable to other types of molecular data as separate modules and is freely available to academic institutions for licensing. PMID:16364175
Internet-based Profiler system as integrative framework to support translational research.
Kim, Robert; Demichelis, Francesca; Tang, Jeffery; Riva, Alberto; Shen, Ronglai; Gibbs, Doug F; Mahavishno, Vasudeva; Chinnaiyan, Arul M; Rubin, Mark A
2005-12-19
Translational research requires taking basic science observations and developing them into clinically useful tests and therapeutics. We have developed a process to develop molecular biomarkers for diagnosis and prognosis by integrating tissue microarray (TMA) technology and an internet-database tool, Profiler. TMA technology allows investigators to study hundreds of patient samples on a single glass slide resulting in the conservation of tissue and the reduction in inter-experimental variability. The Profiler system allows investigator to reliably track, store, and evaluate TMA experiments. Here within we describe the process that has evolved through an empirical basis over the past 5 years at two academic institutions. The generic design of this system makes it compatible with multiple organ system (e.g., prostate, breast, lung, renal, and hematopoietic system,). Studies and folders are restricted to authorized users as required. Over the past 5 years, investigators at 2 academic institutions have scanned 656 TMA experiments and collected 63,311 digital images of these tissue samples. 68 pathologists from 12 major user groups have accessed the system. Two groups directly link clinical data from over 500 patients for immediate access and the remaining groups choose to maintain clinical and pathology data on separate systems. Profiler currently has 170 K data points such as staining intensity, tumor grade, and nuclear size. Due to the relational database structure, analysis can be easily performed on single or multiple TMA experimental results. The TMA module of Profiler can maintain images acquired from multiple systems. We have developed a robust process to develop molecular biomarkers using TMA technology and an internet-based database system to track all steps of this process. This system is extendable to other types of molecular data as separate modules and is freely available to academic institutions for licensing.
Data-driven process decomposition and robust online distributed modelling for large-scale processes
NASA Astrophysics Data System (ADS)
Shu, Zhang; Lijuan, Li; Lijuan, Yao; Shipin, Yang; Tao, Zou
2018-02-01
With the increasing attention of networked control, system decomposition and distributed models show significant importance in the implementation of model-based control strategy. In this paper, a data-driven system decomposition and online distributed subsystem modelling algorithm was proposed for large-scale chemical processes. The key controlled variables are first partitioned by affinity propagation clustering algorithm into several clusters. Each cluster can be regarded as a subsystem. Then the inputs of each subsystem are selected by offline canonical correlation analysis between all process variables and its controlled variables. Process decomposition is then realised after the screening of input and output variables. When the system decomposition is finished, the online subsystem modelling can be carried out by recursively block-wise renewing the samples. The proposed algorithm was applied in the Tennessee Eastman process and the validity was verified.
NASA Astrophysics Data System (ADS)
Pape, Dennis R.
1990-09-01
The present conference discusses topics in optical image processing, optical signal processing, acoustooptic spectrum analyzer systems and components, and optical computing. Attention is given to tradeoffs in nonlinearly recorded matched filters, miniature spatial light modulators, detection and classification using higher-order statistics of optical matched filters, rapid traversal of an image data base using binary synthetic discriminant filters, wideband signal processing for emitter location, an acoustooptic processor for autonomous SAR guidance, and sampling of Fresnel transforms. Also discussed are an acoustooptic RF signal-acquisition system, scanning acoustooptic spectrum analyzers, the effects of aberrations on acoustooptic systems, fast optical digital arithmetic processors, information utilization in analog and digital processing, optical processors for smart structures, and a self-organizing neural network for unsupervised learning.
Amvrazi, Elpiniki G; Albanis, Triantafyllos A
2008-07-23
The processing factors (pesticide concentration found in olive oil/pesticide concentration found in olives) of azinphos methyl, chlorpyrifos, lambda-cyhalothrin, deltamethrin, diazinon, dimethoate, endosulfan, and fenthion were determined in olive oil production process in various laboratory-scale olive oil extractions based on three- or two-phase centrifugation systems in comparison with samples collected during olive oil extractions in conventional olive mills located at different olive oil production areas in Greece. Pesticide analyses were performed using a multiresidue method developed in our laboratory for the determination of different insecticides and herbicides in olive oil by solid-phase extraction techniques coupled to gas chromatography detection (electron capture detection and nitrogen phosphorus detection), optimized, and validated for olive fruits sample preparation. Processing factors were found to vary among the different pesticides studied. Water addition in the oil extraction procedure (as in a three-phase centrifugation system) was found to decrease the processing factors of dimethoate, alpha-endosulfan, diazinon, and chlorpyrifos, whereas those of fenthion, azinphos methyl, beta-endosulfan, lambda-cyhalothrin, and deltamethrin residues were not affected. The water content of olives processed was found to proportionally affect pesticide processing factors. Fenthion sulfoxide and endosulfan sulfate were the major metabolites of fenthion and endosulfan, respectively, that were detected in laboratory-produced olive oils, but only the concentration of fenthion sulfoxide was found to increase with the increase of water addition in the olive oil extraction process.
Analysis of peptides using an integrated microchip HPLC-MS/MS system.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kirby, Brian J.; Chirica, Gabriela S.; Reichmuth, David S.
Hyphendated LC-MS techniques are quickly becoming the standard tool for protemic analyses. For large homogeneous samples, bulk processing methods and capillary injection and separation techniques are suitable. However, for analysis of small or heterogeneous samples, techniques that can manipulate picoliter samples without dilution are required or samples will be lost or corrupted; further, static nanospray-type flowrates are required to maximize SNR. Microchip-level integration of sample injection with separation and mass spectrometry allow small-volume analytes to be processed on chip and immediately injected without dilution for analysis. An on-chip HPLC was fabricated using in situ polymerization of both fixed and mobilemore » polymer monoliths. Integration of the chip with a nanospray MS emitter enables identification of peptides by the use of tandem MS. The chip is capable of analyzing of very small sample volumes (< 200 pl) in short times (< 3 min).« less
An investigation of phase transformation and crystallinity in laser surface modified H13 steel
NASA Astrophysics Data System (ADS)
Aqida, S. N.; Brabazon, D.; Naher, S.
2013-03-01
This paper presents a laser surface modification process of AISI H13 tool steel using 0.09, 0.2 and 0.4 mm size of laser spot with an aim to increase hardness properties. A Rofin DC-015 diffusion-cooled CO2 slab laser was used to process AISI H13 tool steel samples. Samples of 10 mm diameter were sectioned to 100 mm length in order to process a predefined circumferential area. The parameters selected for examination were laser peak power, overlap percentage and pulse repetition frequency (PRF). X-ray diffraction analysis (XRD) was conducted to measure crystallinity of the laser-modified surface. X-ray diffraction patterns of the samples were recorded using a Bruker D8 XRD system with Cu K α ( λ=1.5405 Å) radiation. The diffraction patterns were recorded in the 2 θ range of 20 to 80°. The hardness properties were tested at 981 mN force. The laser-modified surface exhibited reduced crystallinity compared to the un-processed samples. The presence of martensitic phase was detected in the samples processed using 0.4 mm spot size. Though there was reduced crystallinity, a high hardness was measured in the laser-modified surface. Hardness was increased more than 2.5 times compared to the as-received samples. These findings reveal the phase source of the hardening mechanism and grain composition in the laser-modified surface.
High resolution monitoring system for IRE stack releases.
Deconninck, B; De Lellis, C
2013-11-01
The main activity of IRE (Institute for Radio-Element) is radioisotope production of bulk (99)Mo and (131)I for medical application (diagnosis and therapy). Those isotopes are chemically extracted from HEU (High Enriched Uranium) targets activated in reactors. During this process, fission products are released from the targets, including noble gases isotopes (Xe and Kr). Like any nuclear plant, IRE has release limits which are given by the Belgium authority and moreover IRE is in the process of continuously reducing the level of its releases. To achieve this mission, the need of an accurate tool is necessary and IRE has developed a specific monitoring system using a high resolution detector in order to identify and accurately estimate its gaseous releases. This system has a continuous air sampling system in the plant main stack. The sampled gases cross charcoal cartridges where they are slowed down and concentrated for higher detection efficiency. In front of those cartridges is installed an HPGe detector with a detection chain connected to a specific analysis system allowing on-line spectrum analysis. Each isotope can be separately followed without interferences, especially during the production process where high activity can be released. Due to its conception, the system also allows to measure iodine isotopes by integration on the charcoal cartridges. This device is of great help for accurately estimate IRE releases and to help for understanding specific releases and their origin in the production or maintenance process. Copyright © 2013 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Barba Ferrer, Carme; Folch, Albert; Gaju, Núria; Martínez-Alonso, Maira; Carrasquilla, Marc; Grau-Martínez, Alba; Sanchez-Vila, Xavier
2016-04-01
Managed Artificial Recharge (MAR) represents a strategic tool for managing water resources, especially during scarce periods. On one hand, it can increase water stored in aquifers and extract it when weather conditions do not permit exclusive exploitation of surface resources. On the other, it allows improve water quality due the processes occurring into the soil whereas water crosses vadose zone. Barcelona (Catalonia, Spain) conurbation is suffering significant quantitative and qualitative groundwater disturbances. For this reason, Sant Vicenç MAR system, constituted by a sedimentation and an infiltration pond, was constructed in 2009 as the strategic water management infrastructure. Compared with other MAR facilities, this infiltration pond has a reactive bed formed by organic compost and local material. The objective is to promote different redox states allowing more and different degradation of chemical compounds than regular MAR systems. In previous studies in the site, physical and hydrochemical parameters demonstrated that there was indeed a degradation of different pollutants. However, to go a step further understanding the different biogeochemical processes and the related degradation processes occurring in the system, we studied the existing microbial communities. So, molecular techniques were applied in water and soil samples in two different scenarios; the first one, when the system was fully operating and the second when the system was not operating during some months. We have specifically compared microbial diversity and richness indexes and both cluster dendrograms obtained from DGGEs analysis made in each sampling campaign.
Goal Based Testing: A Risk Informed Process
NASA Technical Reports Server (NTRS)
Everline, Chester; Smith, Clayton; Distefano, Sal; Goldin, Natalie
2014-01-01
A process for life demonstration testing is developed, which can reduce the number of resources required by conventional sampling theory while still maintaining the same degree of rigor and confidence level. This process incorporates state-of-the-art probabilistic thinking and is consistent with existing NASA guidance documentation. This view of life testing changes the paradigm of testing a system for many hours to show confidence that a system will last for the required number of years to one that focuses efforts and resources on exploring how the system can fail at end-of-life and building confidence that the failure mechanisms are understood and well mitigated.
Jha, Ashish Kumar
2015-01-01
Glomerular filtration rate (GFR) estimation by plasma sampling method is considered as the gold standard. However, this method is not widely used because the complex technique and cumbersome calculations coupled with the lack of availability of user-friendly software. The routinely used Serum Creatinine method (SrCrM) of GFR estimation also requires the use of online calculators which cannot be used without internet access. We have developed user-friendly software "GFR estimation software" which gives the options to estimate GFR by plasma sampling method as well as SrCrM. We have used Microsoft Windows(®) as operating system and Visual Basic 6.0 as the front end and Microsoft Access(®) as database tool to develop this software. We have used Russell's formula for GFR calculation by plasma sampling method. GFR calculations using serum creatinine have been done using MIRD, Cockcroft-Gault method, Schwartz method, and Counahan-Barratt methods. The developed software is performing mathematical calculations correctly and is user-friendly. This software also enables storage and easy retrieval of the raw data, patient's information and calculated GFR for further processing and comparison. This is user-friendly software to calculate the GFR by various plasma sampling method and blood parameter. This software is also a good system for storing the raw and processed data for future analysis.
Hazard Control Extensions in a COTS Based Data Handling System
NASA Astrophysics Data System (ADS)
Vogel, Torsten; Rakers, Sven; Gronowski, Matthias; Schneegans, Joachim
2011-08-01
EML is an electromagnetic levitator for containerless processing of conductive samples on the International Space Station. This material sciences experiment is running in the European Drawer Rack (EDR) facility. The objective of this experiment is to gain insight into the parameters of liquid metal samples and their crystallisation processes without the influence of container walls. To this end the samples are electromagnetically positioned in a coil system and then heated up beyond their melting point in an ultraclean environment.The EML programme is currently under development by Astrium Space Transportation in Friedrichshafen and Bremen; jointly funded by ESA and DLR (on behalf of BMWi, contract 50WP0808). EML consists of four main modules listed in Table 1. The paper focuses mainly on the architecture and design of the ECM module and its contribution to a safe operation of the experiment. The ECM is a computer system that integrates the power supply to the EML experiment, control functions and video handling and compression features. Experiment control is performed by either telecommand or the execution of predefined experiment scripts.
Tugulea, Anca-Maria; Aranda-Rodriguez, Rocio; Bérubé, Denis; Giddings, Michèle; Lemieux, France; Hnatiw, Joan; Dabeka, Luda; Breton, François
2018-03-01
The National Survey of Disinfection By-Products and Selected Emerging Contaminants investigated the formation of various disinfection by-products and contaminants in 65 water treatment systems (WTSs) across Canada. Results for six iodo-trihalomethanes (iodo-THMs) are reported in this paper. The participating water treatment systems included large, medium and small systems using water sources and treatment processes which were representative of Canadian drinking water. Five water samples (source water, treated water and three water samples along the distribution system) were collected from each treatment system, both under winter and summer conditions. Samples were stabilized, shipped cold and analysed for six iodo-THMs (dichloroiodomethane-DCIM; dibromoiodomethane-DBIM; bromochloroiodomethane-BCIM; chlorodiiodomethane-CDIM; bromodiiodomethane-BDIM and triiodomethane or iodoform-TIM), using a SPME-GC-ECD method developed in our laboratory (MDLs from 0.02 μg/L for iodoform to 0.06 μg/L for bromodiiodomethane). Concentrations of relevant precursors like dissolved organic carbon (DOC), bromide, iodide and total iodine, as well as other water quality parameters, were also determined. Detailed information about the treatment process used at each location was recorded using a questionnaire. The survey showed that one or more iodo-THMs were detected at 31 out of 64 water treatment systems (WTSs) under winter conditions and in 46 out of 64 WTSs under summer conditions (analytical results from one site were excluded due to sampling challenges). Total iodo-THM concentrations measured during this survey ranged from 0.02 μg/L to 21.66 μg/L. The highest total iodo-THM concentration was measured in WTS 63 where all six iodo-THMs were detected and iodoform was present in the highest concentration. The highest iodo-THM formation was found to occur in treatment systems where water sources had naturally occurring ammonium as well as high bromide, high iodide and/or total iodine concentrations. In two such water systems the total concentration of iodo-THMs exceeded the concentration of regulated THMs. Copyright © 2017. Published by Elsevier Ltd.
Kieninger, M; Zech, N; Mulzer, Y; Bele, S; Seemann, M; Künzig, H; Schneiker, A; Gruber, M
2015-05-01
Point of care testing with blood gas analysis (BGA) is an important factor for intensive care medicine. Continuous efforts to optimize workflow, improve safety for the staff and avoid preanalytical mistakes are important and should reflect quality management standards. In a prospective observational study it was investigated whether the implementation of a new system for BGA using labeled syringes and automated processing of the specimens leads to improvements compared to the previously used procedure. In a 4-week test period the time until receiving the final results of the BGA with the standard method used in the clinical routine (control group) was compared to the results in a second 4-week test period using the new labeled syringes and automated processing of the specimens (intervention group). In addition, preanalytical mistakes with both systems were checked during routine daily use. Finally, it was investigated whether a delay of 10 min between taking and analyzing the blood samples alters the results of the BGA. Preanalytical errors were frequently observed in the control group where non-deaerated samples were recorded in 87.3 % but in the intervention group almost all samples (98.9 %) were correctly deaerated. Insufficient homogenization due to omission of manual pivoting was seen in 83.2 % in the control group and in 89.9 % in the intervention group; however, in the intervention group the samples were homogenized automatically during the further analytical process. Although a survey among the staff revealed a high acceptance of the new system and a subjective improvement of workflow, a measurable gain in time after conversion to the new procedure could not be seen. The mean time needed for a complete analysis process until receiving the final results was 244 s in the intervention group and 201 s in the control group. A 10-min delay between taking and analyzing the blood samples led to a significant and clinically relevant elevation of the values for partial pressure of oxygen (pO2) in both groups compared to the results when analyzing the samples immediately (118.4 vs. 148.6 mmHg in the control group and 115.3 vs. 123.7 mmHg in the intervention group). When using standard syringes the partial pressure of carbon dioxide (pCO2) was significantly lower (40.5 vs. 38.3 mmHg) whereas no alterations were seen when using the labeled syringes. The implementation of a new BGA system with labeled syringes and automated processing of the specimens was possible without any difficulties under daily clinical routine conditions in this 10-bed intensive care unit (ICU). A gain of time could not be measured but a reduction in preanalytical errors using the labeled syringes with automated processing was found. Delayed analysis of blood samples can lead to significant changes in pO2 and pCO2 depending on the type of syringe used.
Smart Optical Material Characterization System and Method
NASA Technical Reports Server (NTRS)
Choi, Sang Hyouk (Inventor); Park, Yeonjoon (Inventor)
2015-01-01
Disclosed is a system and method for characterizing optical materials, using steps and equipment for generating a coherent laser light, filtering the light to remove high order spatial components, collecting the filtered light and forming a parallel light beam, splitting the parallel beam into a first direction and a second direction wherein the parallel beam travelling in the second direction travels toward the material sample so that the parallel beam passes through the sample, applying various physical quantities to the sample, reflecting the beam travelling in the first direction to produce a first reflected beam, reflecting the beam that passes through the sample to produce a second reflected beam that travels back through the sample, combining the second reflected beam after it travels back though the sample with the first reflected beam, sensing the light beam produced by combining the first and second reflected beams, and processing the sensed beam to determine sample characteristics and properties.
Inferring the parameters of a Markov process from snapshots of the steady state
NASA Astrophysics Data System (ADS)
Dettmer, Simon L.; Berg, Johannes
2018-02-01
We seek to infer the parameters of an ergodic Markov process from samples taken independently from the steady state. Our focus is on non-equilibrium processes, where the steady state is not described by the Boltzmann measure, but is generally unknown and hard to compute, which prevents the application of established equilibrium inference methods. We propose a quantity we call propagator likelihood, which takes on the role of the likelihood in equilibrium processes. This propagator likelihood is based on fictitious transitions between those configurations of the system which occur in the samples. The propagator likelihood can be derived by minimising the relative entropy between the empirical distribution and a distribution generated by propagating the empirical distribution forward in time. Maximising the propagator likelihood leads to an efficient reconstruction of the parameters of the underlying model in different systems, both with discrete configurations and with continuous configurations. We apply the method to non-equilibrium models from statistical physics and theoretical biology, including the asymmetric simple exclusion process (ASEP), the kinetic Ising model, and replicator dynamics.
Planetary image conversion task
NASA Technical Reports Server (NTRS)
Martin, M. D.; Stanley, C. L.; Laughlin, G.
1985-01-01
The Planetary Image Conversion Task group processed 12,500 magnetic tapes containing raw imaging data from JPL planetary missions and produced an image data base in consistent format on 1200 fully packed 6250-bpi tapes. The output tapes will remain at JPL. A copy of the entire tape set was delivered to US Geological Survey, Flagstaff, Ariz. A secondary task converted computer datalogs, which had been stored in project specific MARK IV File Management System data types and structures, to flat-file, text format that is processable on any modern computer system. The conversion processing took place at JPL's Image Processing Laboratory on an IBM 370-158 with existing software modified slightly to meet the needs of the conversion task. More than 99% of the original digital image data was successfully recovered by the conversion task. However, processing data tapes recorded before 1975 was destructive. This discovery is of critical importance to facilities responsible for maintaining digital archives since normal periodic random sampling techniques would be unlikely to detect this phenomenon, and entire data sets could be wiped out in the act of generating seemingly positive sampling results. Reccomended follow-on activities are also included.
Van Berkel, Gary J.; Kertesz, Vilmos; Orcutt, Matt; ...
2017-11-07
The aim of this work was to demonstrate and to evaluate the analytical performance of a combined falling drop/open port sampling interface (OPSI) system as a simple noncontact, no-carryover, automated system for flow injection analysis with mass spectrometry. The falling sample drops were introduced into the OPSI using a widely available autosampler platform utilizing low cost disposable pipet tips and conventional disposable microtiter well plates. The volume of the drops that fell onto the OPSI was in the 7–15 μL range with an injected sample volume of several hundred nanoliters. Sample drop height, positioning of the internal capillary on themore » sampling end of the probe, and carrier solvent flow rate were optimized for maximum signal. Sample throughput, signal reproducibility, matrix effects, and quantitative analysis capability of the system were established using the drug molecule propranolol and its isotope labeled internal standard in water, unprocessed river water and two commercially available buffer matrices. A sample-to-sample throughput of ~45 s with a ~4.5 s base-to-base flow injection peak profile was obtained in these experiments. In addition, quantitation with minimally processed rat plasma samples was demonstrated with three different statin drugs (atorvastatin, rosuvastatin, and fluvastatin). Direct characterization capability of unprocessed samples was demonstrated by the analysis of neat vegetable oils. Employing the autosampler system for spatially resolved liquid extraction surface sampling exemplified by the analysis of propranolol and its hydroxypropranolol glucuronide phase II metabolites from a rat thin tissue section was also illustrated.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Van Berkel, Gary J.; Kertesz, Vilmos; Orcutt, Matt
The aim of this work was to demonstrate and to evaluate the analytical performance of a combined falling drop/open port sampling interface (OPSI) system as a simple noncontact, no-carryover, automated system for flow injection analysis with mass spectrometry. The falling sample drops were introduced into the OPSI using a widely available autosampler platform utilizing low cost disposable pipet tips and conventional disposable microtiter well plates. The volume of the drops that fell onto the OPSI was in the 7–15 μL range with an injected sample volume of several hundred nanoliters. Sample drop height, positioning of the internal capillary on themore » sampling end of the probe, and carrier solvent flow rate were optimized for maximum signal. Sample throughput, signal reproducibility, matrix effects, and quantitative analysis capability of the system were established using the drug molecule propranolol and its isotope labeled internal standard in water, unprocessed river water and two commercially available buffer matrices. A sample-to-sample throughput of ~45 s with a ~4.5 s base-to-base flow injection peak profile was obtained in these experiments. In addition, quantitation with minimally processed rat plasma samples was demonstrated with three different statin drugs (atorvastatin, rosuvastatin, and fluvastatin). Direct characterization capability of unprocessed samples was demonstrated by the analysis of neat vegetable oils. Employing the autosampler system for spatially resolved liquid extraction surface sampling exemplified by the analysis of propranolol and its hydroxypropranolol glucuronide phase II metabolites from a rat thin tissue section was also illustrated.« less
Software for a GPS-Reflection Remote-Sensing System
NASA Technical Reports Server (NTRS)
Lowe, Stephen
2003-01-01
A special-purpose software Global Positioning System (GPS) receiver designed for remote sensing with reflected GPS signals is described in Delay/Doppler-Mapping GPS-Reflection Remote-Sensing System (NPO-30385), which appears elsewhere in this issue of NASA Tech Briefs. The input accepted by this program comprises raw (open-loop) digitized GPS signals sampled at a rate of about 20 MHz. The program processes the data samples to perform the following functions: detection of signals; tracking of phases and delays; mapping of delay, Doppler, and delay/Doppler waveforms; dual-frequency processing; coherent integrations as short as 125 s; decoding of navigation messages; and precise time tagging of observable quantities. The software can perform these functions on all detectable satellite signals without dead time. Open-loop data collected over water, land, or ice and processed by this software can be further processed to extract geophysical information. Possible examples include mean sea height, wind speed and direction, and significant wave height (for observations over the ocean); bistatic-radar terrain images and measures of soil moisture and biomass (for observations over land); and estimates of ice age, thickness, and surface density (for observations over ice).
Canadian macromolecular crystallography facility: a suite of fully automated beamlines.
Grochulski, Pawel; Fodje, Michel; Labiuk, Shaunivan; Gorin, James; Janzen, Kathryn; Berg, Russ
2012-06-01
The Canadian light source is a 2.9 GeV national synchrotron radiation facility located on the University of Saskatchewan campus in Saskatoon. The small-gap in-vacuum undulator illuminated beamline, 08ID-1, together with the bending magnet beamline, 08B1-1, constitute the Canadian Macromolecular Crystallography Facility (CMCF). The CMCF provides service to more than 50 Principal Investigators in Canada and the United States. Up to 25% of the beam time is devoted to commercial users and the general user program is guaranteed up to 55% of the useful beam time through a peer-review process. CMCF staff provides "Mail-In" crystallography service to users with the highest scored proposals. Both beamlines are equipped with very robust end-stations including on-axis visualization systems, Rayonix 300 CCD series detectors and Stanford-type robotic sample auto-mounters. MxDC, an in-house developed beamline control system, is integrated with a data processing module, AutoProcess, allowing full automation of data collection and data processing with minimal human intervention. Sample management and remote monitoring of experiments is enabled through interaction with a Laboratory Information Management System developed at the facility.
HACCP-Based Programs for Preventing Disease and Injury from Premise Plumbing: A Building Consensus
McCoy, William F.; Rosenblatt, Aaron A.
2015-01-01
Thousands of preventable injuries and deaths are annually caused by microbial, chemical and physical hazards from building water systems. Water is processed in buildings before use; this can degrade the quality of the water. Processing steps undertaken on-site in buildings often include conditioning, filtering, storing, heating, cooling, pressure regulation and distribution through fixtures that restrict flow and temperature. Therefore, prevention of disease and injury requires process management. A process management framework for buildings is the hazard analysis and critical control point (HACCP) adaptation of failure mode effects analysis (FMEA). It has been proven effective for building water system management. Validation is proof that hazards have been controlled under operating conditions and may include many kinds of evidence including cultures of building water samples to detect and enumerate potentially pathogenic microorganisms. However, results from culture tests are often inappropriately used because the accuracy and precision are not sufficient to support specifications for control limit or action triggers. A reliable negative screen is based on genus-level Polymerase Chain Reaction (PCR) for Legionella in building water systems; however, building water samples with positive results from this test require further analysis by culture methods. PMID:26184325
HACCP-Based Programs for Preventing Disease and Injury from Premise Plumbing: A Building Consensus.
McCoy, William F; Rosenblatt, Aaron A
2015-07-09
Thousands of preventable injuries and deaths are annually caused by microbial, chemical and physical hazards from building water systems. Water is processed in buildings before use; this can degrade the quality of the water. Processing steps undertaken on-site in buildings often include conditioning, filtering, storing, heating, cooling, pressure regulation and distribution through fixtures that restrict flow and temperature. Therefore, prevention of disease and injury requires process management. A process management framework for buildings is the hazard analysis and critical control point (HACCP) adaptation of failure mode effects analysis (FMEA). It has been proven effective for building water system management. Validation is proof that hazards have been controlled under operating conditions and may include many kinds of evidence including cultures of building water samples to detect and enumerate potentially pathogenic microorganisms. However, results from culture tests are often inappropriately used because the accuracy and precision are not sufficient to support specifications for control limit or action triggers. A reliable negative screen is based on genus-level Polymerase Chain Reaction (PCR) for Legionella in building water systems; however, building water samples with positive results from this test require further analysis by culture methods.
Enrichment of Thorium (Th) and Lead (Pb) in the early Galaxy
NASA Astrophysics Data System (ADS)
Aoki, Wako; Honda, Satoshi
2010-03-01
We have been determining abundances of Th, Pb and other neutron-capture elements in metal-deficient cool giant stars to constrain the enrichment of heavy elements by the r- and s-processes. Our current sample covers the metallicity range between [Fe/H] = -2.5 and -1.0. (1) The abundance ratios of Pb/Fe and Pb/Eu of most of our stars are approximately constant, and no increase of these ratios with increasing metallicity is found. This result suggests that the Pb abundances of our sample are determined by the r-process with no or little contribution of the s-process. (2) The Th/Eu abundance ratios of our sample show no significant scatter, and the average is lower by 0.2 dex in the logarithmic scale than the solar-system value. This result indicates that the actinides production by the r-process does not show large dispersion, even though r-process models suggest high sensitivity of the actinides production to the nucleosynthesis environment.
Smart Vest: wearable multi-parameter remote physiological monitoring system.
Pandian, P S; Mohanavelu, K; Safeer, K P; Kotresh, T M; Shakunthala, D T; Gopal, Parvati; Padaki, V C
2008-05-01
The wearable physiological monitoring system is a washable shirt, which uses an array of sensors connected to a central processing unit with firmware for continuously monitoring physiological signals. The data collected can be correlated to produce an overall picture of the wearer's health. In this paper, we discuss the wearable physiological monitoring system called 'Smart Vest'. The Smart Vest consists of a comfortable to wear vest with sensors integrated for monitoring physiological parameters, wearable data acquisition and processing hardware and remote monitoring station. The wearable data acquisition system is designed using microcontroller and interfaced with wireless communication and global positioning system (GPS) modules. The physiological signals monitored are electrocardiogram (ECG), photoplethysmogram (PPG), body temperature, blood pressure, galvanic skin response (GSR) and heart rate. The acquired physiological signals are sampled at 250samples/s, digitized at 12-bit resolution and transmitted wireless to a remote physiological monitoring station along with the geo-location of the wearer. The paper describes a prototype Smart Vest system used for remote monitoring of physiological parameters and the clinical validation of the data are also presented.
The External Quality Assessment Scheme (EQAS): Experiences of a medium sized accredited laboratory.
Bhat, Vivek; Chavan, Preeti; Naresh, Chital; Poladia, Pratik
2015-06-15
We put forth our experiences of EQAS, analyzed the result discrepancies, reviewed the corrective actions and also put forth strategies for risk identification and prevention of potential errors in a medical laboratory. For hematology, EQAS samples - blood, peripheral and reticulocyte smears - were received quarterly every year. All the blood samples were processed on HMX hematology analyzer by Beckman-Coulter. For clinical chemistry, lyophilized samples were received and were processed on Siemens Dimension Xpand and RXL analyzers. For microbiology, EQAS samples were received quarterly every year as lyophilized strains along with smears and serological samples. In hematology no outliers were noted for reticulocyte and peripheral smear examination. Only one outlier was noted for CBC. In clinical chemistry outliers (SDI ≥ 2) were noted in 7 samples (23 parameters) out of total 36 samples (756 parameters) processed. Thirteen of these parameters were analyzed as random errors, 3 as transcriptional errors and seven instances of systemic error were noted. In microbiology, one discrepancy was noted in isolate identification and in the grading of smears for AFB by Ziehl Neelsen stain. EQAS along with IQC is a very important tool for maintaining optimal quality of services. Copyright © 2015 Elsevier B.V. All rights reserved.
Tack, Lois C; Thomas, Michelle; Reich, Karl
2007-03-01
Forensic labs globally face the same problem-a growing need to process a greater number and wider variety of samples for DNA analysis. The same forensic lab can be tasked all at once with processing mixed casework samples from crime scenes, convicted offender samples for database entry, and tissue from tsunami victims for identification. Besides flexibility in the robotic system chosen for forensic automation, there is a need, for each sample type, to develop new methodology that is not only faster but also more reliable than past procedures. FTA is a chemical treatment of paper, unique to Whatman Bioscience, and is used for the stabilization and storage of biological samples. Here, the authors describe optimization of the Whatman FTA Purification Kit protocol for use with the AmpFlSTR Identifiler PCR Amplification Kit.
A statistical comparison of two carbon fiber/epoxy fabrication techniques
NASA Technical Reports Server (NTRS)
Hodge, A. J.
1991-01-01
A statistical comparison of the compression strengths of specimens that were fabricated by either a platen press or an autoclave were performed on IM6/3501-6 carbon/epoxy composites of 16-ply (0,+45,90,-45)(sub S2) lay-up configuration. The samples were cured with the same parameters and processing materials. It was found that the autoclaved panels were thicker than the platen press cured samples. Two hundred samples of each type of cure process were compression tested. The autoclaved samples had an average strength of 450 MPa (65.5 ksi), while the press cured samples had an average strength of 370 MPa (54.0 ksi). A Weibull analysis of the data showed that there is only a 30 pct. probability that the two types of cure systems yield specimens that can be considered from the same family.
Mars Sample Return: Mars Ascent Vehicle Mission and Technology Requirements
NASA Technical Reports Server (NTRS)
Bowles, Jeffrey V.; Huynh, Loc C.; Hawke, Veronica M.; Jiang, Xun J.
2013-01-01
A Mars Sample Return mission is the highest priority science mission for the next decade recommended by the recent Decadal Survey of Planetary Science, the key community input process that guides NASAs science missions. A feasibility study was conducted of a potentially simple and low cost approach to Mars Sample Return mission enabled by the use of developing commercial capabilities. Previous studies of MSR have shown that landing an all up sample return mission with a high mass capacity lander is a cost effective approach. The approach proposed is the use of an emerging commercially available capsule to land the launch vehicle system that would return samples to Earth. This paper describes the mission and technology requirements impact on the launch vehicle system design, referred to as the Mars Ascent Vehicle (MAV).
Mars Sample Return: Mars Ascent Vehicle Mission and Technology Requirements
NASA Technical Reports Server (NTRS)
Bowles, Jeffrey V.; Huynh, Loc C.; Hawke, Veronica M.
2013-01-01
A Mars Sample Return mission is the highest priority science mission for the next decade recommended by the recent Decadal Survey of Planetary Science, the key community input process that guides NASA's science missions. A feasibility study was conducted of a potentially simple and low cost approach to Mars Sample Return mission enabled by the use of new commercial capabilities. Previous studies of MSR have shown that landing an all up sample return mission with a high mass capacity lander is a cost effective approach. The approach proposed is the use of a SpaceX Dragon capsule to land the launch vehicle system that would return samples to Earth. This paper describes the mission and technology requirements impact on the launch vehicle system design, referred to as the Mars Ascent Vehicle (MAV).
Preparation of water and ice samples for 39Ar dating by atom trap trace analysis (ATTA)
NASA Astrophysics Data System (ADS)
Schwefel, R.; Reichel, T.; Aeschbach-Hertig, W.; Wagenbach, D.
2012-04-01
Atom trap trace analysis (ATTA) is a new and promising method to measure very rare noble gas radioisotopes in the environment. The applicability of this method for the dating of very old groundwater with 81Kr has already been demonstrated [1]. Recent developments now show its feasibility also for the analysis of 39Ar [2,3], which is an ideal dating tracer for the age range between 50 and 1000 years. This range is of interest in the fields of hydro(geo)logy, oceanography, and glaciology. We present preparation (gas extraction and Ar separation) methods for groundwater and ice samples for later analysis by the ATTA technique. For groundwater, the sample size is less of a limitation than for applications in oceanography or glaciology. Large samples are furthermore needed to enable a comparison with the classical method of 39Ar detection by low-level counting. Therefore, a system was built that enables gas extraction from several thousand liters of water using membrane contactors. This system provides degassing efficiencies greater than 80 % and has successfully been tested in the field. Gas samples are further processed to separate a pure Ar fraction by a gas-chromatographic method based on Li-LSX zeolite as selective adsorber material at very low temperatures. The gas separation achieved by this system is controlled by a quadrupole mass spectrometer. It has successfully been tested and used on real samples. The separation efficiency was found to be strongly temperature dependent in the range of -118 to -130 °C. Since ATTA should enable the analysis of 39Ar on samples of less than 1 ccSTP of Ar (corresponding to about 100 ml of air, 2.5 l of water or 1 kg of ice), a method to separate Ar from small amounts of gas was developed. Titanium sponge was found to absorb 60 ccSTP of reactive gases per g of the getter material with reasonably high absorption rates at high operating temperatures (~ 800 ° C). Good separation (higher than 92 % Ar content in residual gas) was achieved by this gettering process. The other main remaining component is H2, which can be further reduced by operating the Ti getter at lower temperature. Furthermore, a system was designed to degas ice samples, followed by Ar separation by gettering. Ice from an alpine glacier was successfully processed on this system.
Pal, Shyamali
2017-12-01
The presence of Macro prolactin is a significant cause of elevated prolactin resulting in misdiagnosis in all automated systems. Poly ethylene glycol (PEG) pretreatment is the preventive process but such process includes the probability of loss of a fraction of bioactive prolactin. Surprisingly, PEG treated EQAS & IQAS samples in Cobas e 411 are found out to be correlating with direct results of at least 3 immunoassay systems and treated and untreated Cobas e 411 results are comparable by a correlation coefficient. Comparison of EQAS, IQAS and patient samples were done to find out the trueness of such correlation factor. Study with patient's results have established the correlation coefficient is valid for very small concentration of prolactin also. EQAS, IQAS and 150 patient samples were treated with PEG and prolactin results of treated and untreated samples obtained from Roche Cobas e 411. 25 patient's results (treated) were compared with direct results in Advia Centaur, Architect I & Access2 systems. Correlation coefficient was obtained from trend line of the treated and untreated results. Two tailed p-value obtained from regression coefficient(r) and sample size. The correlation coefficient is in the range (0.761-0.771). Reverse correlation range is (1.289-1.301). r value of two sets of calculated results were 0.995. Two tailed p- value is zero approving dismissal of null hypothesis. The z-score of EQAS does not always assure authenticity of resultsPEG precipitation is correlated by the factor 0.761 even in very small concentrationsAbbreviationsGFCgel filtration chromatographyPEGpolyethylene glycolEQASexternal quality assurance systemM-PRLmacro prolactinPRLprolactinECLIAelectro-chemiluminescence immunoassayCLIAclinical laboratory improvement amendmentsIQASinternal quality assurance systemrregression coefficient.
Reliability analysis in the Office of Safety, Environmental, and Mission Assurance (OSEMA)
NASA Astrophysics Data System (ADS)
Kauffmann, Paul J.
1994-12-01
The technical personnel in the SEMA office are working to provide the highest degree of value-added activities to their support of the NASA Langley Research Center mission. Management perceives that reliability analysis tools and an understanding of a comprehensive systems approach to reliability will be a foundation of this change process. Since the office is involved in a broad range of activities supporting space mission projects and operating activities (such as wind tunnels and facilities), it was not clear what reliability tools the office should be familiar with and how these tools could serve as a flexible knowledge base for organizational growth. Interviews and discussions with the office personnel (both technicians and engineers) revealed that job responsibilities ranged from incoming inspection to component or system analysis to safety and risk. It was apparent that a broad base in applied probability and reliability along with tools for practical application was required by the office. A series of ten class sessions with a duration of two hours each was organized and scheduled. Hand-out materials were developed and practical examples based on the type of work performed by the office personnel were included. Topics covered were: Reliability Systems - a broad system oriented approach to reliability; Probability Distributions - discrete and continuous distributions; Sampling and Confidence Intervals - random sampling and sampling plans; Data Analysis and Estimation - Model selection and parameter estimates; and Reliability Tools - block diagrams, fault trees, event trees, FMEA. In the future, this information will be used to review and assess existing equipment and processes from a reliability system perspective. An analysis of incoming materials sampling plans was also completed. This study looked at the issues associated with Mil Std 105 and changes for a zero defect acceptance sampling plan.
Reliability analysis in the Office of Safety, Environmental, and Mission Assurance (OSEMA)
NASA Technical Reports Server (NTRS)
Kauffmann, Paul J.
1994-01-01
The technical personnel in the SEMA office are working to provide the highest degree of value-added activities to their support of the NASA Langley Research Center mission. Management perceives that reliability analysis tools and an understanding of a comprehensive systems approach to reliability will be a foundation of this change process. Since the office is involved in a broad range of activities supporting space mission projects and operating activities (such as wind tunnels and facilities), it was not clear what reliability tools the office should be familiar with and how these tools could serve as a flexible knowledge base for organizational growth. Interviews and discussions with the office personnel (both technicians and engineers) revealed that job responsibilities ranged from incoming inspection to component or system analysis to safety and risk. It was apparent that a broad base in applied probability and reliability along with tools for practical application was required by the office. A series of ten class sessions with a duration of two hours each was organized and scheduled. Hand-out materials were developed and practical examples based on the type of work performed by the office personnel were included. Topics covered were: Reliability Systems - a broad system oriented approach to reliability; Probability Distributions - discrete and continuous distributions; Sampling and Confidence Intervals - random sampling and sampling plans; Data Analysis and Estimation - Model selection and parameter estimates; and Reliability Tools - block diagrams, fault trees, event trees, FMEA. In the future, this information will be used to review and assess existing equipment and processes from a reliability system perspective. An analysis of incoming materials sampling plans was also completed. This study looked at the issues associated with Mil Std 105 and changes for a zero defect acceptance sampling plan.
NASA Technical Reports Server (NTRS)
Brand, R. R.; Barker, J. L.
1983-01-01
A multistage sampling procedure using image processing, geographical information systems, and analytical photogrammetry is presented which can be used to guide the collection of representative, high-resolution spectra and discrete reflectance targets for future satellite sensors. The procedure is general and can be adapted to characterize areas as small as minor watersheds and as large as multistate regions. Beginning with a user-determined study area, successive reductions in size and spectral variation are performed using image analysis techniques on data from the Multispectral Scanner, orbital and simulated Thematic Mapper, low altitude photography synchronized with the simulator, and associated digital data. An integrated image-based geographical information system supports processing requirements.
Continuous flow electrophoresis system experiments on shuttle flights STS-6 and STS-7
NASA Technical Reports Server (NTRS)
Snyder, Robert S.; Rhodes, Percy H.; Miller, Teresa Y.
1987-01-01
A space continuous flow electrophoresis system (CFES) was developed that would incorporate specific modifications to laboratory instruments to take advantage of weightlessness. The specific objectives were to use a model sample material at a high concentration to evaluate the continuous flow electrophoresis process in the CFES instrument and compare its separation resolution and sample throughput with related devices on Earth and to expand the basic knowledge of the limitations imposed by fluid flows and particle concentration effects on the electrophoresis process by careful design and evaluation of the space experiment. Hemoglobin and polysaccharide were selected as primary samples. The results from space show a large band spread of the high concentration of the single species of hemoglobin that was due to the mismatch of electrical conductivity between the sample and the buffer. On STS-7 the major objective was to evaluate the influence of the electrical properties of the sample constituents on the resolution of the CFES. As expected, the polystyrene latex microspheres dispersed in a solution with 3 times the electrical conductivity of the curtain buffer separated with a larger band spread than in the 2nd experiment.
Coarsening in Solid-Liquid Mixtures Studied on the Space Shuttle
NASA Technical Reports Server (NTRS)
Caruso, John J.
1999-01-01
Ostwald ripening, or coarsening, is a process in which large particles in a two-phase mixture grow at the expense of small particles. It is a ubiquitous natural phenomena occurring in the late stages of virtually all phase separation processes. In addition, a large number of commercially important alloys undergo coarsening because they are composed of particles embedded in a matrix. Many of them, such as high-temperature superalloys used for turbine blade materials and low-temperature aluminum alloys, coarsen in the solid state. In addition, many alloys, such as the tungsten-heavy metal systems, coarsen in the solid-liquid state during liquid phase sintering. Numerous theories have been proposed that predict the rate at which the coarsening process occurs and the shape of the particle size distribution. Unfortunately, these theories have never been tested using a system that satisfies all the assumptions of the theory. In an effort to test these theories, NASA studied the coarsening process in a solid-liquid mixture composed of solid tin particles in a liquid lead-tin matrix. On Earth, the solid tin particles float to the surface of the sample, like ice in water. In contrast, in a microgravity environment this does not occur. The microstructures in the ground- and space-processed samples (see the photos) show clearly the effects of gravity on the coarsening process. The STS-83-processed sample (right image) shows nearly spherical uniformly dispersed solid tin particles. In contrast, the identically processed, ground-based sample (left image) shows significant density-driven, nonspherical particles, and because of the higher effective solid volume fraction, a larger particle size after the same coarsening time. The "Coarsening in Solid-Liquid Mixtures" (CSLM) experiment was conducted in the Middeck Glovebox facility (MGBX) flown aboard the shuttle in the Microgravity Science Laboratory (MSL-1/1R) on STS-83/94. The primary objective of CSLM is to measure the temporal evolution of the solid particles during coarsening.
Isotope pattern deconvolution as rising tool for isotope tracer studies in environmental research
NASA Astrophysics Data System (ADS)
Irrgeher, Johanna; Zitek, Andreas; Prohaska, Thomas
2014-05-01
During the last decade stable isotope tracers have emerged as versatile tool in ecological research. Besides 'intrinsic' isotope tracers caused by the natural variation of isotopes, the intentional introduction of 'extrinsic' enriched stable isotope tracers into biological systems has gained significant interest. Hereby the induced change in the natural isotopic composition of an element allows amongst others for studying the fate and fluxes of metals, trace elements and species in organisms or provides an intrinsic marker or tag of particular biological samples. Due to the shoreless potential of this methodology, the number of publications dealing with applications of isotope (double) spikes as tracers to address research questions in 'real world systems' is constantly increasing. However, some isotope systems like the natural Sr isotopic system, although potentially very powerful for this type of application, are still rarely used, mainly because their adequate measurement/determination poses major analytical challenges; as e.g. Sr is available in significant amounts in natural samples. In addition, biological systems underlie complex processes such as metabolism, adsorption/desorption or oxidation/reduction. As a consequence, classic evaluation approaches such as the isotope dilution mass spectrometry equation are often not applicable because of the unknown amount of tracer finally present in the sample. Isotope pattern deconvolution (IPD), based on multiple linear regression, serves as simplified alternative data processing strategy to double spike isotope dilution calculations. The outstanding advantage of this mathematical tool lies in the possibility of deconvolving the isotope pattern in a spiked sample without knowing the quantities of enriched isotope tracer being incorporated into the natural sample matrix as well as the degree of impurities and species-interconversion (e.g. from sample preparation). Here, the potential of IPD for environmental tracer studies is critically discussed, where special emphasis is set on evaluating different data processing strategies on the example of enriched stable Sr isotopes.1 The analytical key parameters such as blank (Kr, Sr and Rb), variation of the natural Sr isotopic composition in the sample, mass bias, interferences (Rb) and total combined uncertainty are considered. A full metrological protocol for data processing using IPD is presented based on data gained during two transgenerational marking studies of fish, where the transfer of a Sr isotope double spike (84Sr and 86Sr) from female spawners of common carp (Cyprinus carpio L.) and brown trout (Salmo trutta f.f.)2 to the centre of the otoliths of their offspring was studied by (LA)-MC-ICP-MS. 1J. Irrgeher, A. Zitek, M. Cervicek and T. Prohaska, J. Anal. At. Spectrom., 2014, 29, 193-200. 2A. Zitek, J. Irrgeher, M. Kletzl, T. Weismann and T. Prohaska, Fish. Manage. Ecol., 2013, 20, 654-361.
Remote laser drilling and sampling system for the detection of concealed explosives
NASA Astrophysics Data System (ADS)
Wild, D.; Pschyklenk, L.; Theiß, C.; Holl, G.
2017-05-01
The detection of hazardous materials like explosives is a central issue in national security in the field of counterterrorism. One major task includes the development of new methods and sensor systems for the detection. Many existing remote or standoff methods like infrared or raman spectroscopy find their limits, if the hazardous material is concealed in an object. Imaging technologies using x-ray or terahertz radiation usually yield no information about the chemical content itself. However, the exact knowledge of the real threat potential of a suspicious object is crucial for disarming the device. A new approach deals with a laser drilling and sampling system for the use as verification detector for suspicious objects. Central part of the system is a miniaturised, diode pumped Nd:YAG laser oscillator-amplifier. The system allows drilling into most materials like metals, synthetics or textiles with bore hole diameters in the micron scale. During the drilling process, the hazardous material can be sampled for further investigation with suitable detection methods. In the reported work, laser induced breakdown spectroscopy (LIBS) is used to monitor the drilling process and to classify the drilled material. Also experiments were carried out to show the system's ability to not ignite even sensitive explosives like triacetone triperoxide (TATP). The detection of concealed hazardous material is shown for different explosives using liquid chromatography and ion mobility spectrometry.
NASA Astrophysics Data System (ADS)
Laskar, S.; Bordoloi, S.
2016-01-01
This paper presents an instrumentation system to measure the degradation in lubricating oil using a bare, tapered and bent multi-mode optical fiber (BTBMOF) sensor probe and a temperature probe. The sensor system consists of (i) a bare, tapered and bent multi-mode optical fiber (BTBMOF) as optical sensor along with a laser source and a LDR (Light Dependent Resistor) as detector (ii) a temperature sensor (iii) a ATmega microcontroller based data acquisition system and (iv) a trained ANN for processing and calibration. The BTBMOF sensor and the temperature sensor are used to provide the measure of refractive index (RI) and the temperature of a lubricating oil sample. A microcontroller based instrumentation system with trained ANN algorithm has been developed to determine the degradation of the lubricating oil sample by sampling the readings of the optical fiber sensor, and the temperature sensor.
Hasar, U C
2009-05-01
A microcontroller-based noncontact and nondestructive microwave free-space measurement system for real-time and dynamic determination of complex permittivity of lossy liquid materials has been proposed. The system is comprised of two main sections--microwave and electronic. While the microwave section provides for measuring only the amplitudes of reflection coefficients, the electronic section processes these data and determines the complex permittivity using a general purpose microcontroller. The proposed method eliminates elaborate liquid sample holder preparation and only requires microwave components to perform reflection measurements from one side of the holder. In addition, it explicitly determines the permittivity of lossy liquid samples from reflection measurements at different frequencies without any knowledge on sample thickness. In order to reduce systematic errors in the system, we propose a simple calibration technique, which employs simple and readily available standards. The measurement system can be a good candidate for industrial-based applications.
Shackleton, David; Pagram, Jenny; Ives, Lesley; Vanhinsbergh, Des
2018-06-02
The RapidHIT™ 200 System is a fully automated sample-to-DNA profile system designed to produce high quality DNA profiles within 2h. The use of RapidHIT™ 200 System within the United Kingdom Criminal Justice System (UKCJS) has required extensive development and validation of methods with a focus on AmpFℓSTR ® NGMSElect™ Express PCR kit to comply with specific regulations for loading to the UK National DNA Database (NDNAD). These studies have been carried out using single source reference samples to simulate live reference samples taken from arrestees and victims for elimination. The studies have shown that the system is capable of generating high quality profile and has achieved the accreditations necessary to load to the NDNAD; a first for the UK. Copyright © 2018 Elsevier B.V. All rights reserved.
Development of a System Model for Non-Invasive Quantification of Bilirubin in Jaundice Patients
NASA Astrophysics Data System (ADS)
Alla, Suresh K.
Neonatal jaundice is a medical condition which occurs in newborns as a result of an imbalance between the production and elimination of bilirubin. Excess bilirubin in the blood stream diffuses into the surrounding tissue leading to a yellowing of the skin. An optical system integrated with a signal processing system is used as a platform to noninvasively quantify bilirubin concentration through the measurement of diffuse skin reflectance. Initial studies have lead to the generation of a clinical analytical model for neonatal jaundice which generates spectral reflectance data for jaundiced skin with varying levels of bilirubin concentration in the tissue. The spectral database built using the clinical analytical model is then used as a test database to validate the signal processing system in real time. This evaluation forms the basis for understanding the translation of this research to human trials. The clinical analytical model and signal processing system have been successful validated on three spectral databases. First spectral database is constructed using a porcine model as a surrogate for neonatal skin tissue. Samples of pig skin were soaked in bilirubin solutions of varying concentrations to simulate jaundice skin conditions. The resulting skins samples were analyzed with our skin reflectance systems producing bilirubin concentration values that show a high correlation (R2 = 0.94) to concentration of the bilirubin solution that each porcine tissue sample is soaked in. The second spectral database is the spectral measurements collected on human volunteers to quantify the different chromophores and other physical properties of the tissue such a Hematocrit, Hemoglobin etc. The third spectral database is the spectral data collected at different time periods from the moment a bruise is induced.
Fast and automatic thermographic material identification for the recycling process
NASA Astrophysics Data System (ADS)
Haferkamp, Heinz; Burmester, Ingo
1998-03-01
Within the framework of the future closed loop recycling process the automatic and economical sorting of plastics is a decisive element. The at the present time available identification and sorting systems are not yet suitable for the sorting of technical plastics since essential demands, as the realization of high recognition reliability and identification rates considering the variety of technical plastics, can not be guaranteed. Therefore the Laser Zentrum Hannover e.V. in cooperation with the Hoerotron GmbH and the Preussag Noell GmbH has carried out investigations on a rapid thermographic and laser-supported material- identification-system for automatic material-sorting- systems. The automatic identification of different engineering plastics coming from electronic or automotive waste is possible. Identification rates up to 10 parts per second are allowed by the effort from fast IR line scanners. The procedure is based on the following principle: within a few milliseconds a spot on the relevant sample is heated by a CO2 laser. The samples different and specific chemical and physical material properties cause different temperature distributions on their surfaces that are measured by a fast IR-linescan system. This 'thermal impulse response' has to be analyzed by means of a computer system. Investigations have shown that it is possible to analyze more than 18 different sorts of plastics at a frequency of 10 Hz. Crucial for the development of such a system is the rapid processing of imaging data, the minimization of interferences caused by oscillating samples geometries, and a wide range of possible additives in plastics in question. One possible application area is sorting of plastics coming from car- and electronic waste recycling.
Frequency division multiplexed multi-color fluorescence microscope system
NASA Astrophysics Data System (ADS)
Le, Vu Nam; Yang, Huai Dong; Zhang, Si Chun; Zhang, Xin Rong; Jin, Guo Fan
2017-10-01
Grayscale camera can only obtain gray scale image of object, while the multicolor imaging technology can obtain the color information to distinguish the sample structures which have the same shapes but in different colors. In fluorescence microscopy, the current method of multicolor imaging are flawed. Problem of these method is affecting the efficiency of fluorescence imaging, reducing the sampling rate of CCD etc. In this paper, we propose a novel multiple color fluorescence microscopy imaging method which based on the Frequency division multiplexing (FDM) technology, by modulating the excitation lights and demodulating the fluorescence signal in frequency domain. This method uses periodic functions with different frequency to modulate amplitude of each excitation lights, and then combine these beams for illumination in a fluorescence microscopy imaging system. The imaging system will detect a multicolor fluorescence image by a grayscale camera. During the data processing, the signal obtained by each pixel of the camera will be processed with discrete Fourier transform, decomposed by color in the frequency domain and then used inverse discrete Fourier transform. After using this process for signals from all of the pixels, monochrome images of each color on the image plane can be obtained and multicolor image is also acquired. Based on this method, this paper has constructed and set up a two-color fluorescence microscope system with two excitation wavelengths of 488 nm and 639 nm. By using this system to observe the linearly movement of two kinds of fluorescent microspheres, after the data processing, we obtain a two-color fluorescence dynamic video which is consistent with the original image. This experiment shows that the dynamic phenomenon of multicolor fluorescent biological samples can be generally observed by this method. Compared with the current methods, this method can obtain the image signals of each color at the same time, and the color video's frame rate is consistent with the frame rate of the camera. The optical system is simpler and does not need extra color separation element. In addition, this method has a good filtering effect on the ambient light or other light signals which are not affected by the modulation process.
Dynamic modeling and parameter estimation of a radial and loop type distribution system network
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jun Qui; Heng Chen; Girgis, A.A.
1993-05-01
This paper presents a new identification approach to three-phase power system modeling and model reduction taking power system network as multi-input, multi-output (MIMO) processes. The model estimate can be obtained in discrete-time input-output form, discrete- or continuous-time state-space variable form, or frequency-domain impedance transfer function matrix form. An algorithm for determining the model structure of this MIMO process is described. The effect of measurement noise on the approach is also discussed. This approach has been applied on a sample system and simulation results are also presented in this paper.
Tank Monitoring and Document control System (TMACS) As Built Software Design Document
DOE Office of Scientific and Technical Information (OSTI.GOV)
GLASSCOCK, J.A.
This document describes the software design for the Tank Monitor and Control System (TMACS). This document captures the existing as-built design of TMACS as of November 1999. It will be used as a reference document to the system maintainers who will be maintaining and modifying the TMACS functions as necessary. The heart of the TMACS system is the ''point-processing'' functionality where a sample value is received from the field sensors and the value is analyzed, logged, or alarmed as required. This Software Design Document focuses on the point-processing functions.
The validation of forensic DNA extraction systems to utilize soil contaminated biological evidence.
Kasu, Mohaimin; Shires, Karen
2015-07-01
The production of full DNA profiles from biological evidence found in soil has a high failure rate due largely to the inhibitory substance humic acid (HA). Abundant in various natural soils, HA co-extracts with DNA during extraction and inhibits DNA profiling by binding to the molecular components of the genotyping assay. To successfully utilize traces of soil contaminated evidence, such as that found at many murder and rape crime scenes in South Africa, a reliable HA removal extraction system would often be selected based on previous validation studies. However, for many standard forensic DNA extraction systems, peer-reviewed publications detailing the efficacy on soil evidence is either lacking or is incomplete. Consequently, these sample types are often not collected or fail to yield suitable DNA material due to the use of unsuitable methodology. The aim of this study was to validate the common forensic DNA collection and extraction systems used in South Africa, namely DNA IQ, FTA elute and Nucleosave for processing blood and saliva contaminated with HA. A forensic appropriate volume of biological evidence was spiked with HA (0, 0.5, 1.5 and 2.5 mg/ml) and processed through each extraction protocol for the evaluation of HA removal using QPCR and STR-genotyping. The DNA IQ magnetic bead system effectively removed HA from highly contaminated blood and saliva, and generated consistently acceptable STR profiles from both artificially spiked samples and crude soil samples. This system is highly recommended for use on soil-contaminated evidence over the cellulose card-based systems currently being preferentially used for DNA sample collection. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
A Microbial Assessment Scheme to measure microbial performance of Food Safety Management Systems.
Jacxsens, L; Kussaga, J; Luning, P A; Van der Spiegel, M; Devlieghere, F; Uyttendaele, M
2009-08-31
A Food Safety Management System (FSMS) implemented in a food processing industry is based on Good Hygienic Practices (GHP), Hazard Analysis Critical Control Point (HACCP) principles and should address both food safety control and assurance activities in order to guarantee food safety. One of the most emerging challenges is to assess the performance of a present FSMS. The objective of this work is to explain the development of a Microbial Assessment Scheme (MAS) as a tool for a systematic analysis of microbial counts in order to assess the current microbial performance of an implemented FSMS. It is assumed that low numbers of microorganisms and small variations in microbial counts indicate an effective FSMS. The MAS is a procedure that defines the identification of critical sampling locations, the selection of microbiological parameters, the assessment of sampling frequency, the selection of sampling method and method of analysis, and finally data processing and interpretation. Based on the MAS assessment, microbial safety level profiles can be derived, indicating which microorganisms and to what extent they contribute to food safety for a specific food processing company. The MAS concept is illustrated with a case study in the pork processing industry, where ready-to-eat meat products are produced (cured, cooked ham and cured, dried bacon).
Flow Cytometry: Impact on Early Drug Discovery.
Edwards, Bruce S; Sklar, Larry A
2015-07-01
Modern flow cytometers can make optical measurements of 10 or more parameters per cell at tens of thousands of cells per second and more than five orders of magnitude dynamic range. Although flow cytometry is used in most drug discovery stages, "sip-and-spit" sampling technology has restricted it to low-sample-throughput applications. The advent of HyperCyt sampling technology has recently made possible primary screening applications in which tens of thousands of compounds are analyzed per day. Target-multiplexing methodologies in combination with extended multiparameter analyses enable profiling of lead candidates early in the discovery process, when the greatest numbers of candidates are available for evaluation. The ability to sample small volumes with negligible waste reduces reagent costs, compound usage, and consumption of cells. Improved compound library formatting strategies can further extend primary screening opportunities when samples are scarce. Dozens of targets have been screened in 384- and 1536-well assay formats, predominantly in academic screening lab settings. In concert with commercial platform evolution and trending drug discovery strategies, HyperCyt-based systems are now finding their way into mainstream screening labs. Recent advances in flow-based imaging, mass spectrometry, and parallel sample processing promise dramatically expanded single-cell profiling capabilities to bolster systems-level approaches to drug discovery. © 2015 Society for Laboratory Automation and Screening.
Flow Cytometry: Impact On Early Drug Discovery
Edwards, Bruce S.; Sklar, Larry A.
2015-01-01
Summary Modern flow cytometers can make optical measurements of 10 or more parameters per cell at tens-of-thousands of cells per second and over five orders of magnitude dynamic range. Although flow cytometry is used in most drug discovery stages, “sip-and-spit” sampling technology has restricted it to low sample throughput applications. The advent of HyperCyt sampling technology has recently made possible primary screening applications in which tens-of-thousands of compounds are analyzed per day. Target-multiplexing methodologies in combination with extended multi-parameter analyses enable profiling of lead candidates early in the discovery process, when the greatest numbers of candidates are available for evaluation. The ability to sample small volumes with negligible waste reduces reagent costs, compound usage and consumption of cells. Improved compound library formatting strategies can further extend primary screening opportunities when samples are scarce. Dozens of targets have been screened in 384- and 1536-well assay formats, predominantly in academic screening lab settings. In concert with commercial platform evolution and trending drug discovery strategies, HyperCyt-based systems are now finding their way into mainstream screening labs. Recent advances in flow-based imaging, mass spectrometry and parallel sample processing promise dramatically expanded single cell profiling capabilities to bolster systems level approaches to drug discovery. PMID:25805180
Fleischer, Heidi; Ramani, Kinjal; Blitti, Koffi; Roddelkopf, Thomas; Warkentin, Mareike; Behrend, Detlef; Thurow, Kerstin
2018-02-01
Automation systems are well established in industries and life science laboratories, especially in bioscreening and high-throughput applications. An increasing demand of automation solutions can be seen in the field of analytical measurement in chemical synthesis, quality control, and medical and pharmaceutical fields, as well as research and development. In this study, an automation solution was developed and optimized for the investigation of new biliary endoprostheses (stents), which should reduce clogging after implantation in the human body. The material inside the stents (incrustations) has to be controlled regularly and under identical conditions. The elemental composition is one criterion to be monitored in stent development. The manual procedure was transferred to an automated process including sample preparation, elemental analysis using inductively coupled plasma mass spectrometry (ICP-MS), and data evaluation. Due to safety issues, microwave-assisted acid digestion was executed outside of the automation system. The performance of the automated process was determined and validated. The measurement results and the processing times were compared for both the manual and the automated procedure. Finally, real samples of stent incrustations and pig bile were analyzed using the automation system.
Lifetime of Feshbach dimers in a Fermi-Fermi mixture of 6Li and 40K
NASA Astrophysics Data System (ADS)
Jag, M.; Cetina, M.; Lous, R. S.; Grimm, R.; Levinsen, J.; Petrov, D. S.
2016-12-01
We present a joint experimental and theoretical investigation of the lifetime of weakly bound dimers formed near narrow interspecies Feshbach resonances in mass-imbalanced Fermi-Fermi systems, considering the specific example of a mixture of 6Li and 40K atoms. Our work addresses the central question of the increase in the stability of the dimers resulting from Pauli suppression of collisional losses, which is a well-known effect in mass-balanced fermionic systems near broad resonances. We present measurements of the spontaneous dissociation of dimers in dilute samples, and of the collisional losses in dense samples arising from both dimer-dimer processes and from atom-dimer processes. We find that all loss processes are suppressed close to the Feshbach resonance. Our general theoretical approach for fermionic mixtures near narrow Feshbach resonances provides predictions for the suppression of collisional decay as a function of the detuning from resonance, and we find excellent agreement with the experimental benchmarks provided by our 40K-6Li system. We finally present model calculations for other Feshbach-resonant Fermi-Fermi systems, which are of interest for experiments in the near future.
Lu, Zhan; Zhang, Jianyi; Xu, Lizhou; Li, Yanbin; Chen, Siyu; Ye, Zunzhong; Wang, Jianping
2017-01-01
A simple, highly-automated instrument system used for on-site detection of foodborne pathogens based on fluorescence was designed, fabricated, and preliminarily tested in this paper. A corresponding method has been proved effective in our previous studies. This system utilizes a light-emitting diode (LED) to excite fluorescent labels and a spectrometer to record the fluorescence signal from samples. A rotation stage for positioning and switching samples was innovatively designed for high-throughput detection, ten at most in one single run. We also developed software based on LabVIEW for data receiving, processing, and the control of the whole system. In the test of using a pure quantum dot (QD) solution as a standard sample, detection results from this home-made system were highly-relevant with that from a well-commercialized product and even slightly better reproducibility was found. And in the test of three typical kinds of food-borne pathogens, fluorescence signals recorded by this system are highly proportional to the variation of the sample concentration, with a satisfied limit of detection (LOD) (nearly 102–103 CFU·mL−1 in food samples). Additionally, this instrument system is low-cost and easy-to-use, showing a promising potential for on-site rapid detection of food-borne pathogens. PMID:28241478
Lu, Zhan; Zhang, Jianyi; Xu, Lizhou; Li, Yanbin; Chen, Siyu; Ye, Zunzhong; Wang, Jianping
2017-02-23
A simple, highly-automated instrument system used for on-site detection of foodborne pathogens based on fluorescence was designed, fabricated, and preliminarily tested in this paper. A corresponding method has been proved effective in our previous studies. This system utilizes a light-emitting diode (LED) to excite fluorescent labels and a spectrometer to record the fluorescence signal from samples. A rotation stage for positioning and switching samples was innovatively designed for high-throughput detection, ten at most in one single run. We also developed software based on LabVIEW for data receiving, processing, and the control of the whole system. In the test of using a pure quantum dot (QD) solution as a standard sample, detection results from this home-made system were highly-relevant with that from a well-commercialized product and even slightly better reproducibility was found. And in the test of three typical kinds of food-borne pathogens, fluorescence signals recorded by this system are highly proportional to the variation of the sample concentration, with a satisfied limit of detection (LOD) (nearly 10²-10³ CFU·mL -1 in food samples). Additionally, this instrument system is low-cost and easy-to-use, showing a promising potential for on-site rapid detection of food-borne pathogens.
NASA Astrophysics Data System (ADS)
Saprykin, A. A.; Sharkeev, Yu P.; Ibragimov, E. A.; Babakova, E. V.; Dudikhin, D. V.
2016-07-01
Alloys based on the titanium-niobium system are widely used in implant production. It is conditional, first of all, on the low modulus of elasticity and bio-inert properties of an alloy. These alloys are especially important for tooth replacement and orthopedic surgery. At present alloys based on the titanium-niobium system are produced mainly using conventional metallurgical methods. The further subtractive manufacturing an end product results in a lot of wastes, increasing, therefore, its cost. The alternative of these processes is additive manufacturing. Selective laser melting is a technology, which makes it possible to synthesize products of metal powders and their blends. The point of this technology is laser melting a layer of a powdered material; then a sintered layer is coated with the next layer of powder etc. Complex products and working prototypes are made on the base of this technology. The authors of this paper address to the issue of applying selective laser melting in order to synthesize a binary alloy of a composite powder based on the titanium-niobium system. A set of 10x10 mm samples is made in various process conditions. The samples are made by an experimental selective laser synthesis machine «VARISKAF-100MB». The machine provides adjustment of the following process variables: laser emission power, scanning rate and pitch, temperature of powder pre-heating, thickness of the layer to be sprinkled, and diameter of laser spot focusing. All samples are made in the preliminary vacuumized shielding atmosphere of argon. The porosity and thickness of the sintered layer related to the laser emission power are shown at various scanning rates. It is revealed that scanning rate and laser emission power are adjustable process variables, having the greatest effect on forming the sintered layer.
Viscosity Meaurement Technique for Metal Fuels
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ban, Heng; Kennedy, Rory
2015-02-09
Metallic fuels have exceptional transient behavior, excellent thermal conductivity, and a more straightforward reprocessing path, which does not separate out pure plutonium from the process stream. Fabrication of fuel containing minor actinides and rare earth (RE) elements for irradiation tests, for instance, U-20Pu-3Am-2Np-1.0RE-15Zr samples at the Idaho National Laboratory, is generally done by melt casting in an inert atmosphere. For the design of a casting system and further scale up development, computational modeling of the casting process is needed to provide information on melt flow and solidification for process optimization. Therefore, there is a need for melt viscosity data, themore » most important melt property that controls the melt flow. The goal of the project was to develop a measurement technique that uses fully sealed melt sample with no Americium vapor loss to determine the viscosity of metallic melts and at temperatures relevant to the casting process. The specific objectives of the project were to: develop mathematical models to establish the principle of the measurement method, design and build a viscosity measurement prototype system based on the established principle, and calibrate the system and quantify the uncertainty range. The result of the project indicates that the oscillation cup technique is applicable for melt viscosity measurement. Detailed mathematical models of innovative sample ampoule designs were developed to not only determine melt viscosity, but also melt density under certain designs. Measurement uncertainties were analyzed and quantified. The result of this project can be used as the initial step toward the eventual goal of establishing a viscosity measurement system for radioactive melts.« less
Curating NASA's Past, Present, and Future Astromaterial Sample Collections
NASA Technical Reports Server (NTRS)
Zeigler, R. A.; Allton, J. H.; Evans, C. A.; Fries, M. D.; McCubbin, F. M.; Nakamura-Messenger, K.; Righter, K.; Zolensky, M.; Stansbery, E. K.
2016-01-01
The Astromaterials Acquisition and Curation Office at NASA Johnson Space Center (hereafter JSC curation) is responsible for curating all of NASA's extraterrestrial samples. JSC presently curates 9 different astromaterials collections in seven different clean-room suites: (1) Apollo Samples (ISO (International Standards Organization) class 6 + 7); (2) Antarctic Meteorites (ISO 6 + 7); (3) Cosmic Dust Particles (ISO 5); (4) Microparticle Impact Collection (ISO 7; formerly called Space-Exposed Hardware); (5) Genesis Solar Wind Atoms (ISO 4); (6) Stardust Comet Particles (ISO 5); (7) Stardust Interstellar Particles (ISO 5); (8) Hayabusa Asteroid Particles (ISO 5); (9) OSIRIS-REx Spacecraft Coupons and Witness Plates (ISO 7). Additional cleanrooms are currently being planned to house samples from two new collections, Hayabusa 2 (2021) and OSIRIS-REx (2023). In addition to the labs that house the samples, we maintain a wide variety of infra-structure facilities required to support the clean rooms: HEPA-filtered air-handling systems, ultrapure dry gaseous nitrogen systems, an ultrapure water system, and cleaning facilities to provide clean tools and equipment for the labs. We also have sample preparation facilities for making thin sections, microtome sections, and even focused ion-beam sections. We routinely monitor the cleanliness of our clean rooms and infrastructure systems, including measurements of inorganic or organic contamination, weekly airborne particle counts, compositional and isotopic monitoring of liquid N2 deliveries, and daily UPW system monitoring. In addition to the physical maintenance of the samples, we track within our databases the current and ever changing characteristics (weight, location, etc.) of more than 250,000 individually numbered samples across our various collections, as well as more than 100,000 images, and countless "analog" records that record the sample processing records of each individual sample. JSC Curation is co-located with JSC's Astromaterials Research Office, which houses a world-class suite of analytical instrumentation and scientists. We leverage these labs and personnel to better curate the samples. Part of the cu-ration process is planning for the future, and we refer to these planning efforts as "advanced curation". Advanced Curation is tasked with developing procedures, technology, and data sets necessary for curating new types of collections as envi-sioned by NASA exploration goals. We are (and have been) planning for future cu-ration, including cold curation, extended curation of ices and volatiles, curation of samples with special chemical considerations such as perchlorate-rich samples, and curation of organically- and biologically-sensitive samples.
Factors to Consider in Designing Aerosol Inlet Systems for Engine Exhaust Plume Sampling
NASA Technical Reports Server (NTRS)
Anderson, Bruce
2004-01-01
This document consists of viewgraphs of charts and diagrams of considerations to take when sampling the engine exhaust plume. It includes a chart that compares the emissions from various fuels, a diagram and charts of the various processes and conditions that influence the particulate size and concentration,
Dynamic Speckle Imaging with Low-Cost Devices
ERIC Educational Resources Information Center
Vannoni, Maurizio; Trivi, Marcelo; Arizaga, Ricardo; Rabal, Hector; Molesini, Giuseppe
2008-01-01
Light from a rough sample surface illuminated with a laser consists of a speckle pattern. If the surface evolves with time, the pattern becomes dynamic, following the activity of the sample. This phenomenon is used both in research and in industry to monitor processes and systems that change with time. The measuring equipment generally includes…
Biotechnology Science Experiments on Mir
NASA Technical Reports Server (NTRS)
Kroes, Roger L.
1999-01-01
This paper describes the microgravity biotechnology experiments carried out on the Shuttle/Mir program. Four experiments investigated the growth of protein crystals, and three investigated cellular growth. Many hundreds of protein samples were processed using four different techniques. The objective of these experiments was to determine optimum conditions for the growth of very high quality single crystals to be used for structure determination. The Biotechnology System (BTS) was used to process the three cell growth investigations. The samples processed by these experiments were: bovine chondrocytes, human renal epithelial cells, and human breast cancer cells and endothelial cells. The objective was to determine the unique properties of cell aggregates produced in the microgravity environment.
Three-level sampler having automated thresholds
NASA Technical Reports Server (NTRS)
Jurgens, R. F.
1976-01-01
A three-level sampler is described that has its thresholds controlled automatically so as to track changes in the statistics of the random process being sampled. In particular, the mean value is removed and the ratio of the standard deviation of the random process to the threshold is maintained constant. The system is configured in such a manner that slow drifts in the level comparators and digital-to-analog converters are also removed. The ratio of the standard deviation to threshold level may be chosen within the constraints of the ratios of two integers N and M. These may be chosen to minimize the quantizing noise of the sampled process.
Bauer, Daniel R; Otter, Michael; Chafin, David R
2018-01-01
Studying and developing preanalytical tools and technologies for the purpose of obtaining high-quality samples for histological assays is a growing field. Currently, there does not exist a standard practice for collecting, fixing, and monitoring these precious samples. There has been some advancement in standardizing collection for the highest profile tumor types, such as breast, where HER2 testing drives therapeutic decisions. This review examines the area of tissue collection, transport, and monitoring of formalin diffusion and details a prototype system that could be used to help standardize tissue collection efforts. We have surveyed recent primary literature sources and conducted several site visits to understand the most error-prone processes in histology laboratories. This effort identified errors that resulted from sample collection techniques and subsequent transport delays from the operating room (OR) to the histology laboratories. We have therefore devised a prototype sample collection and transport concept. The system consists of a custom data logger and cold transport box and takes advantage of a novel cold + warm (named 2 + 2) fixation method. This review highlights the beneficial aspects of standardizing tissue collection, fixation, and monitoring. In addition, a prototype system is introduced that could help standardize these processes and is compatible with use directly in the OR and from remote sites.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wright, R.M.; Zander, M.E.; Brown, S.K.
1992-09-01
This paper describes the application of video image processing to beam profile measurements on the Ground Test Accelerator (GTA). A diagnostic was needed to measure beam profiles in the intermediate matching section (IMS) between the radio-frequency quadrupole (RFQ) and the drift tube linac (DTL). Beam profiles are measured by injecting puffs of gas into the beam. The light emitted from the beam-gas interaction is captured and processed by a video image processing system, generating the beam profile data. A general purpose, modular and flexible video image processing system, imagetool, was used for the GTA image profile measurement. The development ofmore » both software and hardware for imagetool and its integration with the GTA control system (GTACS) will be discussed. The software includes specialized algorithms for analyzing data and calibrating the system. The underlying design philosophy of imagetool was tested by the experience of building and using the system, pointing the way for future improvements. The current status of the system will be illustrated by samples of experimental data.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wright, R.M.; Zander, M.E.; Brown, S.K.
1992-01-01
This paper describes the application of video image processing to beam profile measurements on the Ground Test Accelerator (GTA). A diagnostic was needed to measure beam profiles in the intermediate matching section (IMS) between the radio-frequency quadrupole (RFQ) and the drift tube linac (DTL). Beam profiles are measured by injecting puffs of gas into the beam. The light emitted from the beam-gas interaction is captured and processed by a video image processing system, generating the beam profile data. A general purpose, modular and flexible video image processing system, imagetool, was used for the GTA image profile measurement. The development ofmore » both software and hardware for imagetool and its integration with the GTA control system (GTACS) will be discussed. The software includes specialized algorithms for analyzing data and calibrating the system. The underlying design philosophy of imagetool was tested by the experience of building and using the system, pointing the way for future improvements. The current status of the system will be illustrated by samples of experimental data.« less
[Automatic adjustment control system for DC glow discharge plasma source].
Wan, Zhen-zhen; Wang, Yong-qing; Li, Xiao-jia; Wang, Hai-zhou; Shi, Ning
2011-03-01
There are three important parameters in the DC glow discharge process, the discharge current, discharge voltage and argon pressure in discharge source. These parameters influence each other during glow discharge process. This paper presents an automatic control system for DC glow discharge plasma source. This system collects and controls discharge voltage automatically by adjusting discharge source pressure while the discharge current is constant in the glow discharge process. The design concept, circuit principle and control program of this automatic control system are described. The accuracy is improved by this automatic control system with the method of reducing the complex operations and manual control errors. This system enhances the control accuracy of glow discharge voltage, and reduces the time to reach discharge voltage stability. The glow discharge voltage stability test results with automatic control system are provided as well, the accuracy with automatic control system is better than 1% FS which is improved from 4% FS by manual control. Time to reach discharge voltage stability has been shortened to within 30 s by automatic control from more than 90 s by manual control. Standard samples like middle-low alloy steel and tin bronze have been tested by this automatic control system. The concentration analysis precision has been significantly improved. The RSDs of all the test result are better than 3.5%. In middle-low alloy steel standard sample, the RSD range of concentration test result of Ti, Co and Mn elements is reduced from 3.0%-4.3% by manual control to 1.7%-2.4% by automatic control, and that for S and Mo is also reduced from 5.2%-5.9% to 3.3%-3.5%. In tin bronze standard sample, the RSD range of Sn, Zn and Al elements is reduced from 2.6%-4.4% to 1.0%-2.4%, and that for Si, Ni and Fe is reduced from 6.6%-13.9% to 2.6%-3.5%. The test data is also shown in this paper.
Perera, Rushini S.; Ding, Xavier C.; Tully, Frank; Oliver, James; Bright, Nigel; Bell, David; Chiodini, Peter L.; Gonzalez, Iveth J.; Polley, Spencer D.
2017-01-01
Background Accurate and efficient detection of sub-microscopic malaria infections is crucial for enabling rapid treatment and interruption of transmission. Commercially available malaria LAMP kits have excellent diagnostic performance, though throughput is limited by the need to prepare samples individually. Here, we evaluate the clinical performance of a newly developed high throughput (HTP) sample processing system for use in conjunction with the Eiken malaria LAMP kit. Methods The HTP system utilised dried blood spots (DBS) and liquid whole blood (WB), with parallel sample processing of 94 samples per run. The system was evaluated using 699 samples of known infection status pre-determined by gold standard nested PCR. Results The sensitivity and specificity of WB-HTP-LAMP was 98.6% (95% CI, 95.7–100), and 99.7% (95% CI, 99.2–100); sensitivity of DBS-HTP-LAMP was 97.1% (95% CI, 93.1–100), and specificity 100% against PCR. At parasite densities greater or equal to 2 parasites/μL, WB and DBS HTP-LAMP showed 100% sensitivity and specificity against PCR. At densities less than 2 p/μL, WB-HTP-LAMP sensitivity was 88.9% (95% CI, 77.1–100) and specificity was 99.7% (95% CI, 99.2–100); sensitivity and specificity of DBS-HTP-LAMP was 77.8% (95% CI, 54.3–99.5) and 100% respectively. Conclusions The HTP-LAMP system is a highly sensitive diagnostic test, with the potential to allow large scale population screening in malaria elimination campaigns. PMID:28166235
The design and realization of a three-dimensional video system by means of a CCD array
NASA Astrophysics Data System (ADS)
Boizard, J. L.
1985-12-01
Design features and principles and initial tests of a prototype three-dimensional robot vision system based on a laser source and a CCD detector array is described. The use of a laser as a coherent illumination source permits the determination of the relief using one emitter since the location of the source is a known quantity with low distortion. The CCD signal detector array furnishes an acceptable signal/noise ratio and, when wired to an appropriate signal processing system, furnishes real-time data on the return signals, i.e., the characteristic points of an object being scanned. Signal processing involves integration of 29 kB of data per 100 samples, with sampling occurring at a rate of 5 MHz (the CCDs) and yielding an image every 12 msec. Algorithms for filtering errors from the data stream are discussed.
Van Berkel, Gary J.; Kertesz, Vilmos
2016-11-15
An “Open Access”-like mass spectrometric platform to fully utilize the simplicity of the manual open port sampling interface for rapid characterization of unprocessed samples by liquid introduction atmospheric pressure ionization mass spectrometry has been lacking. The in-house developed integrated software with a simple, small and relatively low-cost mass spectrometry system introduced here fills this void. Software was developed to operate the mass spectrometer, to collect and process mass spectrometric data files, to build a database and to classify samples using such a database. These tasks were accomplished via the vendorprovided software libraries. Sample classification based on spectral comparison utilized themore » spectral contrast angle method. As a result, using the developed software platform near real-time sample classification is exemplified using a series of commercially available blue ink rollerball pens and vegetable oils. In the case of the inks, full scan positive and negative ion ESI mass spectra were both used for database generation and sample classification. For the vegetable oils, full scan positive ion mode APCI mass spectra were recorded. The overall accuracy of the employed spectral contrast angle statistical model was 95.3% and 98% in case of the inks and oils, respectively, using leave-one-out cross-validation. In conclusion, this work illustrates that an open port sampling interface/mass spectrometer combination, with appropriate instrument control and data processing software, is a viable direct liquid extraction sampling and analysis system suitable for the non-expert user and near real-time sample classification via database matching.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Van Berkel, Gary J.; Kertesz, Vilmos
An “Open Access”-like mass spectrometric platform to fully utilize the simplicity of the manual open port sampling interface for rapid characterization of unprocessed samples by liquid introduction atmospheric pressure ionization mass spectrometry has been lacking. The in-house developed integrated software with a simple, small and relatively low-cost mass spectrometry system introduced here fills this void. Software was developed to operate the mass spectrometer, to collect and process mass spectrometric data files, to build a database and to classify samples using such a database. These tasks were accomplished via the vendorprovided software libraries. Sample classification based on spectral comparison utilized themore » spectral contrast angle method. As a result, using the developed software platform near real-time sample classification is exemplified using a series of commercially available blue ink rollerball pens and vegetable oils. In the case of the inks, full scan positive and negative ion ESI mass spectra were both used for database generation and sample classification. For the vegetable oils, full scan positive ion mode APCI mass spectra were recorded. The overall accuracy of the employed spectral contrast angle statistical model was 95.3% and 98% in case of the inks and oils, respectively, using leave-one-out cross-validation. In conclusion, this work illustrates that an open port sampling interface/mass spectrometer combination, with appropriate instrument control and data processing software, is a viable direct liquid extraction sampling and analysis system suitable for the non-expert user and near real-time sample classification via database matching.« less
Baez-Cazull, S. E.; McGuire, J.T.; Cozzarelli, I.M.; Voytek, M.A.
2008-01-01
Determining the processes governing aqueous biogeochemistry in a wetland hydrologically linked to an underlying contaminated aquifer is challenging due to the complex exchange between the systems and their distinct responses to changes in precipitation, recharge, and biological activities. To evaluate temporal and spatial processes in the wetland-aquifer system, water samples were collected using cm-scale multichambered passive diffusion samplers (peepers) to span the wetland-aquifer interface over a period of 3 yr. Samples were analyzed for major cations and anions, methane, and a suite of organic acids resulting in a large dataset of over 8000 points, which was evaluated using multivariate statistics. Principal component analysis (PCA) was chosen with the purpose of exploring the sources of variation in the dataset to expose related variables and provide insight into the biogeochemical processes that control the water chemistry of the system. Factor scores computed from PCA were mapped by date and depth. Patterns observed suggest that (i) fermentation is the process controlling the greatest variability in the dataset and it peaks in May; (ii) iron and sulfate reduction were the dominant terminal electron-accepting processes in the system and were associated with fermentation but had more complex seasonal variability than fermentation; (iii) methanogenesis was also important and associated with bacterial utilization of minerals as a source of electron acceptors (e.g., barite BaSO4); and (iv) seasonal hydrological patterns (wet and dry periods) control the availability of electron acceptors through the reoxidation of reduced iron-sulfur species enhancing iron and sulfate reduction. Copyright ?? 2008 by the American Society of Agronomy, Crop Science Society of America, and Soil Science Society of America. All rights reserved.
A fuzzy classifier system for process control
NASA Technical Reports Server (NTRS)
Karr, C. L.; Phillips, J. C.
1994-01-01
A fuzzy classifier system that discovers rules for controlling a mathematical model of a pH titration system was developed by researchers at the U.S. Bureau of Mines (USBM). Fuzzy classifier systems successfully combine the strengths of learning classifier systems and fuzzy logic controllers. Learning classifier systems resemble familiar production rule-based systems, but they represent their IF-THEN rules by strings of characters rather than in the traditional linguistic terms. Fuzzy logic is a tool that allows for the incorporation of abstract concepts into rule based-systems, thereby allowing the rules to resemble the familiar 'rules-of-thumb' commonly used by humans when solving difficult process control and reasoning problems. Like learning classifier systems, fuzzy classifier systems employ a genetic algorithm to explore and sample new rules for manipulating the problem environment. Like fuzzy logic controllers, fuzzy classifier systems encapsulate knowledge in the form of production rules. The results presented in this paper demonstrate the ability of fuzzy classifier systems to generate a fuzzy logic-based process control system.
Lassahn, Gordon D.; Lancaster, Gregory D.; Apel, William A.; Thompson, Vicki S.
2013-01-08
Image portion identification methods, image parsing methods, image parsing systems, and articles of manufacture are described. According to one embodiment, an image portion identification method includes accessing data regarding an image depicting a plurality of biological substrates corresponding to at least one biological sample and indicating presence of at least one biological indicator within the biological sample and, using processing circuitry, automatically identifying a portion of the image depicting one of the biological substrates but not others of the biological substrates.
Gamma ray spectroscopy monitoring method and apparatus
Stagg, William R; Policke, Timothy A
2017-05-16
The present invention relates generally to the field of gamma ray spectroscopy monitoring and a system for accomplishing same to monitor one or more aspects of various isotope production processes. In one embodiment, the present invention relates to a monitoring system, and method of utilizing same, for monitoring one or more aspects of an isotope production process where the monitoring system comprises: (A) at least one sample cell; (B) at least one measuring port; (C) at least one adjustable collimator device; (D) at least one shutter; and (E) at least one high resolution gamma ray spectrometer.
Dynamically re-configurable CMOS imagers for an active vision system
NASA Technical Reports Server (NTRS)
Yang, Guang (Inventor); Pain, Bedabrata (Inventor)
2005-01-01
A vision system is disclosed. The system includes a pixel array, at least one multi-resolution window operation circuit, and a pixel averaging circuit. The pixel array has an array of pixels configured to receive light signals from an image having at least one tracking target. The multi-resolution window operation circuits are configured to process the image. Each of the multi-resolution window operation circuits processes each tracking target within a particular multi-resolution window. The pixel averaging circuit is configured to sample and average pixels within the particular multi-resolution window.
BROADBAND DIGITAL GEOPHYSICAL TELEMETRY SYSTEM.
Seeley, Robert L.; Daniels, Jeffrey J.
1984-01-01
A system has been developed to simultaneously sample and transmit digital data from five remote geophysical data receiver stations to a control station that processes, displays, and stores the data. A microprocessor in each remote station receives commands from the control station over a single telemetry channel.
Information Science Research: The Search for the Nature of Information.
ERIC Educational Resources Information Center
Kochen, Manfred
1984-01-01
High-level scientific research in the information sciences is illustrated by sampling of recent discoveries involving adaptive information processing strategies, computer and information systems, centroid scaling, economic growth of computer and communication industries, and information flow in biological systems. Relationship of information…
Rodríguez, Rogelio; Avivar, Jessica; Ferrer, Laura; Leal, Luz O; Cerdà, Victor
2012-07-15
A novel lab-on-valve system has been developed for strontium determination in environmental samples. Miniaturized lab-on-valve system potentially offers facilities to allow any kind of chemical and physical processes, including fluidic and microcarrier bead control, homogenous reaction and liquid-solid interaction. A rapid, inexpensive and fully automated method for the separation and preconcentration of total and radioactive strontium, using a solid phase extraction material (Sr-Resin), has been developed. Total strontium concentrations are determined by ICP-OES and (90)Sr activities by a low background proportional counter. The method has been successfully applied to different water samples of environmental interest. The proposed system offers minimization of sample handling, drastic reduction of reagent volume, improvement of the reproducibility and sample throughput and attains a significant decrease of both time and cost per analysis. The LLD of the total Sr reached is 1.8ng and the minimum detectable activity for (90)Sr is 0.008Bq. The repeatability of the separation procedure is 1.2% (n=10). Copyright © 2011 Elsevier B.V. All rights reserved.
Design of signal reception and processing system of embedded ultrasonic endoscope
NASA Astrophysics Data System (ADS)
Li, Ming; Yu, Feng; Zhang, Ruiqiang; Li, Yan; Chen, Xiaodong; Yu, Daoyin
2009-11-01
Embedded Ultrasonic Endoscope, based on embedded microprocessor and embedded real-time operating system, sends a micro ultrasonic probe into coelom through the biopsy channel of the Electronic Endoscope to get the fault histology features of digestive organs by rotary scanning, and acquires the pictures of the alimentary canal mucosal surface. At the same time, ultrasonic signals are processed by signal reception and processing system, forming images of the full histology of the digestive organs. Signal Reception and Processing System is an important component of Embedded Ultrasonic Endoscope. However, the traditional design, using multi-level amplifiers and special digital processing circuits to implement signal reception and processing, is no longer satisfying the standards of high-performance, miniaturization and low power requirements that embedded system requires, and as a result of the high noise that multi-level amplifier brought, the extraction of small signal becomes hard. Therefore, this paper presents a method of signal reception and processing based on double variable gain amplifier and FPGA, increasing the flexibility and dynamic range of the Signal Reception and Processing System, improving system noise level, and reducing power consumption. Finally, we set up the embedded experiment system, using a transducer with the center frequency of 8MHz to scan membrane samples, and display the image of ultrasonic echo reflected by each layer of membrane, with a frame rate of 5Hz, verifying the correctness of the system.
Sutton, Stephen R. [University of Chicago, Chicago, Illinois, United States
2017-12-09
Sample-return missions and natural collection processes have provided us with a surprisingly extensive collection of matter from Solar System bodies other than the Earth. These collections include samples from the Moon, Mars, asteroids, interplanetary dust, and, recently, from the Sun (solar wind) and a comet. This presentation will describe some of these materials, how they were collected, and what we have learned from them. Synchrotron radiation analyses of these materials are playing an increasingly valuable role in unraveling the histories and properities of the parent Solar System bodies.
Uervirojnangkoorn, Monarin; Zeldin, Oliver B.; Lyubimov, Artem Y.; ...
2015-03-17
There is considerable potential for X-ray free electron lasers (XFELs) to enable determination of macromolecular crystal structures that are difficult to solve using current synchrotron sources. Prior XFEL studies often involved the collection of thousands to millions of diffraction images, in part due to limitations of data processing methods. We implemented a data processing system based on classical post-refinement techniques, adapted to specific properties of XFEL diffraction data. When applied to XFEL data from three different proteins collected using various sample delivery systems and XFEL beam parameters, our method improved the quality of the diffraction data as well as themore » resulting refined atomic models and electron density maps. Moreover, the number of observations for a reflection necessary to assemble an accurate data set could be reduced to a few observations. In conclusion, these developments will help expand the applicability of XFEL crystallography to challenging biological systems, including cases where sample is limited.« less
Uervirojnangkoorn, Monarin; Zeldin, Oliver B.; Lyubimov, Artem Y.; ...
2015-03-17
There is considerable potential for X-ray free electron lasers (XFELs) to enable determination of macromolecular crystal structures that are difficult to solve using current synchrotron sources. Prior XFEL studies often involved the collection of thousands to millions of diffraction images, in part due to limitations of data processing methods. We implemented a data processing system based on classical post-refinement techniques, adapted to specific properties of XFEL diffraction data. When applied to XFEL data from three different proteins collected using various sample delivery systems and XFEL beam parameters, our method improved the quality of the diffraction data as well as themore » resulting refined atomic models and electron density maps. Moreover, the number of observations for a reflection necessary to assemble an accurate data set could be reduced to a few observations. These developments will help expand the applicability of XFEL crystallography to challenging biological systems, including cases where sample is limited.« less
Uervirojnangkoorn, Monarin; Zeldin, Oliver B; Lyubimov, Artem Y; Hattne, Johan; Brewster, Aaron S; Sauter, Nicholas K; Brunger, Axel T; Weis, William I
2015-01-01
There is considerable potential for X-ray free electron lasers (XFELs) to enable determination of macromolecular crystal structures that are difficult to solve using current synchrotron sources. Prior XFEL studies often involved the collection of thousands to millions of diffraction images, in part due to limitations of data processing methods. We implemented a data processing system based on classical post-refinement techniques, adapted to specific properties of XFEL diffraction data. When applied to XFEL data from three different proteins collected using various sample delivery systems and XFEL beam parameters, our method improved the quality of the diffraction data as well as the resulting refined atomic models and electron density maps. Moreover, the number of observations for a reflection necessary to assemble an accurate data set could be reduced to a few observations. These developments will help expand the applicability of XFEL crystallography to challenging biological systems, including cases where sample is limited. DOI: http://dx.doi.org/10.7554/eLife.05421.001 PMID:25781634
López Marzo, Adaris M; Pons, Josefina; Blake, Diane A; Merkoçi, Arben
2013-04-02
Nowadays, the development of systems, devices, or methods that integrate several process steps into one multifunctional step for clinical, environmental, or industrial purposes constitutes a challenge for many ongoing research projects. Here, we present a new integrated paper based cadmium (Cd(2+)) immunosensing system in lateral flow format, which integrates the sample treatment process with the analyte detection process. The principle of Cd(2+) detection is based on competitive reaction between the cadmium-ethylenediaminetetraacetic acid-bovine serum albumin-gold nanoparticles (Cd-EDTA-BSA-AuNP) conjugate deposited on the conjugation pad strip and the Cd-EDTA complex formed in the analysis sample for the same binding sites of the 2A81G5 monoclonal antibody (mAb), specific to Cd-EDTA but not Cd(2+) free, which is immobilized onto the test line. This platform operates without any sample pretreatment step for Cd(2+) detection thanks to an extra conjugation pad that ensures Cd(2+) complexation with EDTA and interference masking through ovalbumin (OVA). The detection and quantification limits found for the device were 0.1 and 0.4 ppb, respectively, these being the lowest limits reported up to now for metal sensors based on paper. The accuracy of the device was evaluated by addition of known quantities of Cd(2+) to different drinking water samples and subsequent Cd(2+) content analysis. Sample recoveries ranged from 95 to 105% and the coefficient of variation for the intermediate precision assay was less than 10%. In addition, the results obtained here were compared with those obtained with the well-established inductively coupled plasma emission spectroscopy (ICPES) and the analysis of certificate standard samples.
The Recent Developments in Sample Preparation for Mass Spectrometry-Based Metabolomics.
Gong, Zhi-Gang; Hu, Jing; Wu, Xi; Xu, Yong-Jiang
2017-07-04
Metabolomics is a critical member in systems biology. Although great progress has been achieved in metabolomics, there are still some problems in sample preparation, data processing and data interpretation. In this review, we intend to explore the roles, challenges and trends in sample preparation for mass spectrometry- (MS-) based metabolomics. The newly emerged sample preparation methods were also critically examined, including laser microdissection, in vivo sampling, dried blood spot, microwave, ultrasound and enzyme-assisted extraction, as well as microextraction techniques. Finally, we provide some conclusions and perspectives for sample preparation in MS-based metabolomics.
Method for using polarization gating to measure a scattering sample
Baba, Justin S.
2015-08-04
Described herein are systems, devices, and methods facilitating optical characterization of scattering samples. A polarized optical beam can be directed to pass through a sample to be tested. The optical beam exiting the sample can then be analyzed to determine its degree of polarization, from which other properties of the sample can be determined. In some cases, an apparatus can include a source of an optical beam, an input polarizer, a sample, an output polarizer, and a photodetector. In some cases, a signal from a photodetector can be processed through attenuation, variable offset, and variable gain.
NASA Astrophysics Data System (ADS)
Alexander, Jennifer M.; Grassian, V. H.; Young, M. A.; Kleiber, P. D.
2015-03-01
Visible light scattering phase function and linear polarization profiles of mineral dust components processed with organic acids and humic material are measured, and results are compared to T-matrix simulations of the scattering properties. Processed samples include quartz mixed with humic material, and calcite reacted with acetic and oxalic acids. Clear differences in light scattering properties are observed for all three processed samples when compared to the unprocessed dust or organic salt products. Results for quartz processed with humic acid sodium salt (NaHA) indicate the presence of both internally mixed quartz-NaHA particles and externally mixed NaHA aerosol. Simulations of light scattering suggest that the processed quartz particles become more moderate in shape due to the formation of a coating of humic material over the mineral core. Experimental results for calcite reacted with acetic acid are consistent with an external mixture of calcite and the reaction product, calcium acetate. Modeling of the light scattering properties does not require any significant change to the calcite particle shape distribution although morphology changes cannot be ruled out by our data. It is expected that calcite reacted with oxalic acid will produce internally mixed particles of calcite and calcium oxalate due to the low solubility of the product salt. However, simulations of the scattering for the calcite-oxalic acid system result in rather poor fits to the data when compared to the other samples. The poor fit provides a less accurate picture of the impact of processing in the calcite-oxalic acid system.
Output statistics of laser anemometers in sparsely seeded flows
NASA Technical Reports Server (NTRS)
Edwards, R. V.; Jensen, A. S.
1982-01-01
It is noted that until very recently, research on this topic concentrated on the particle arrival statistics and the influence of the optical parameters on them. Little attention has been paid to the influence of subsequent processing on the measurement statistics. There is also controversy over whether the effects of the particle statistics can be measured. It is shown here that some of the confusion derives from a lack of understanding of the experimental parameters that are to be controlled or known. A rigorous framework is presented for examining the measurement statistics of such systems. To provide examples, two problems are then addressed. The first has to do with a sample and hold processor, the second with what is called a saturable processor. The sample and hold processor converts the output to a continuous signal by holding the last reading until a new one is obtained. The saturable system is one where the maximum processable rate is arrived at by the dead time of some unit in the system. At high particle rates, the processed rate is determined through the dead time.
Study of injection molded microcellular polyamide-6 nanocomposites
Mingjun Yuan; Lih-Sheng Turng; Shaoqin Gong; Daniel Caulfield; Chris Hunt; Rick Spindler
2004-01-01
This study aims to explore the processing benefits and property improvements of combining nanocomposites with microcellular injection molding. The microcellular nanocomposite processing was performed on an injection-molding machine equipped with a commercially available supercritical fluid (SCF) system. The molded samples produced based on the Design of Experiments (...
NASA Technical Reports Server (NTRS)
Dean, Derrick; Haque, Anwar; Vaidya, Uday; Meador, Michael A. (Technical Monitor)
2003-01-01
PMR-type polyimides are regarded as state of the art high temperature polymers, due to their excellent thermo-mechanical properties and thermo-oxidative stability. One of their drawbacks, however is the inability to process them using cost-effective processing methods such as Resin Transfer Molding (RTM) and Resin Film Infusion (RFI). Development of low viscosity, high temperature polymers has been the subject of intense research. Recently, a new generation of low viscosity polyimides were synthesized by the introduction of twisted biphenyl or binapthyl groups into the backbone. This report details the progress for Year 1, which has involved acquiring samples and initiating Phases I and II of the proposed research. Specifically, studies of the process-property relationships of a series of polymers using oligomers based on 2,3,3' ,4'-biphenyltertracarboxylic dianhydride (PBDA) and a mixture of a diamine, BAX and a triamine, 1,3,5-Tris (4-aminophenoxybenzene), TAB, where the amount of TAB was varied have been initiated. The sample containing 10 percent TAB possesses a slightly higher degree of crystalline order versus that of the 20 percent TAB sample, based on x-ray diffraction studies of the b-staged oligomers. Both systems lose all of the crystalline order upon curing, however. The chemorheology has been studied as a function of the TAB content. While the magnitude of the viscosity is essentially the same for both systems, the cure kinetics of the 10 percent TAB system is faster than that for the 20 percent TAB system. The sample exhibits a melting-recrystallization-remelting behavior before the crosslinking commences. Correlation of other kinetic parameters, such as the activation energies for curing, the Tg and mechanical properties to the structure of these systems is underway. Future studies will involve characterization of mechanical and thermal properties of the pure resins and the fabrication of fiber reinforced composites using these materials.
Sample levitation and melt in microgravity
NASA Technical Reports Server (NTRS)
Moynihan, Philip I. (Inventor)
1990-01-01
A system is described for maintaining a sample material in a molten state and away from the walls of a container in a microgravity environment, as in a space vehicle. A plurality of sources of electromagnetic radiation, such as an infrared wavelength, are spaced about the object, with the total net electromagnetic radiation applied to the object being sufficient to maintain it in a molten state, and with the vector sum of the applied radiation being in a direction to maintain the sample close to a predetermined location away from the walls of a container surrounding the sample. For a processing system in a space vehicle that orbits the Earth, the net radiation vector is opposite the velocity of the orbiting vehicle.
Sample levitation and melt in microgravity
NASA Technical Reports Server (NTRS)
Moynihan, Philip I. (Inventor)
1987-01-01
A system is described for maintaining a sample material in a molten state and away from the walls of a container in a microgravity environment, as in a space vehicle. A plurality of sources of electromagnetic radiation, such as of an infrared wavelength, are spaced about the object, with the total net electromagnetic radiation applied to the object being sufficient to maintain it in a molten state, and with the vector sum of the applied radiation being in a direction to maintain the sample close to a predetermined location away from the walls of a container surrounding the sample. For a processing system in a space vehicle that orbits the Earth, the net radiation vector is opposite the velocity of the orbiting vehicle.
A Stochastic Diffusion Process for the Dirichlet Distribution
Bakosi, J.; Ristorcelli, J. R.
2013-03-01
The method of potential solutions of Fokker-Planck equations is used to develop a transport equation for the joint probability ofNcoupled stochastic variables with the Dirichlet distribution as its asymptotic solution. To ensure a bounded sample space, a coupled nonlinear diffusion process is required: the Wiener processes in the equivalent system of stochastic differential equations are multiplicative with coefficients dependent on all the stochastic variables. Individual samples of a discrete ensemble, obtained from the stochastic process, satisfy a unit-sum constraint at all times. The process may be used to represent realizations of a fluctuating ensemble ofNvariables subject to a conservation principle.more » Similar to the multivariate Wright-Fisher process, whose invariant is also Dirichlet, the univariate case yields a process whose invariant is the beta distribution. As a test of the results, Monte Carlo simulations are used to evolve numerical ensembles toward the invariant Dirichlet distribution.« less
The planning and establishment of a sample preparation laboratory for drug discovery
Dufresne, Claude
2000-01-01
Nature has always been a productive source of new drugs. With the advent of high-throughput screening, it has now become possible to rapidly screen large sample collections. In addition to seeking greater diversity from natural product sources (micro-organisms, plants, etc.), fractionation of the crude extracts prior to screening is becoming a more important part of our efforts. As sample preparation protocols become more involved, automation can help to achieve and maintain a desired sample throughput. To address the needs of our screening program, two robotic systems were designed. The first system processes crude extracts all the way to 96-well plates, containing solutions suitable for screening in biological and biochemical assays. The system can dissolve crude extracts, fractionate them on solid-phase extraction cartridges, dry and weigh each fraction, re-dissolve them to a known concentration, and prepare mother plates. The second system replicates mother plates into a number of daughter plates. PMID:18924691
Use of an Electronic Tongue System and Fuzzy Logic to Analyze Water Samples
NASA Astrophysics Data System (ADS)
Braga, Guilherme S.; Paterno, Leonardo G.; Fonseca, Fernando J.
2009-05-01
An electronic tongue (ET) system incorporating 8 chemical sensors was used in combination with two pattern recognition tools, namely principal component analysis (PCA) and Fuzzy logic for discriminating/classification of water samples from different sources (tap, distilled and three brands of mineral water). The Fuzzy program exhibited a higher accuracy than the PCA and allowed the ET to classify correctly 4 in 5 types of water. Exception was made for one brand of mineral water which was sometimes misclassified as tap water. On the other hand, the PCA grouped water samples in three clusters, one with the distilled water; a second with tap water and one brand of mineral water, and the third with the other two other brands of mineral water. Samples in the second and third clusters could not be distinguished. Nevertheless, close grouping between repeated tests indicated that the ET system response is reproducible. The potential use of the Fuzzy logic as the data processing tool in combination with an electronic tongue system is discussed.
Low cost light-sheet microscopy for whole brain imaging
NASA Astrophysics Data System (ADS)
Kumar, Manish; Nasenbeny, Jordan; Kozorovitskiy, Yevgenia
2018-02-01
Light-sheet microscopy has evolved as an indispensable tool in imaging biological samples. It can image 3D samples at fast speed, with high-resolution optical sectioning, and with reduced photobleaching effects. These properties make light-sheet microscopy ideal for imaging fluorophores in a variety of biological samples and organisms, e.g. zebrafish, drosophila, cleared mouse brains, etc. While most commercial turnkey light-sheet systems are expensive, the existing lower cost implementations, e.g. OpenSPIM, are focused on achieving high-resolution imaging of small samples or organisms like zebrafish. In this work, we substantially reduce the cost of light-sheet microscope system while targeting to image much larger samples, i.e. cleared mouse brains, at single-cell resolution. The expensive components of a lightsheet system - excitation laser, water-immersion objectives, and translation stage - are replaced with an incoherent laser diode, dry objectives, and a custom-built Arduino-controlled translation stage. A low-cost CUBIC protocol is used to clear fixed mouse brain samples. The open-source platforms of μManager and Fiji support image acquisition, processing, and visualization. Our system can easily be extended to multi-color light-sheet microscopy.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Santamaria, L.; Siller, H. R.; Garcia-Ortiz, C. E., E-mail: cegarcia@cicese.mx
In this work, we present an alternative optical method to determine the probe-sample separation distance in a scanning near-field optical microscope. The experimental method is based in a Lloyd’s mirror interferometer and offers a measurement precision deviation of ∼100 nm using digital image processing and numerical analysis. The technique can also be strategically combined with the characterization of piezoelectric actuators and stability evaluation of the optical system. It also opens the possibility for the development of an automatic approximation control system valid for probe-sample distances from 5 to 500 μm.
Implementation of in-line infrared monitor in full-scale anaerobic digestion process.
Spanjers, H; Bouvier, J C; Steenweg, P; Bisschops, I; van Gils, W; Versprille, B
2006-01-01
During start up but also during normal operation, anaerobic reactor systems should be run and monitored carefully to secure trouble-free operation, because the process is vulnerable to disturbances such as temporary overloading, biomass wash out and influent toxicity. The present method of monitoring is usually by manual sampling and subsequent laboratory analysis. Data collection, processing and feedback to system operation is manual and ad hoc, and involves high-level operator skills and attention. As a result, systems tend to be designed at relatively conservative design loading rates resulting in significant over-sizing of reactors and thus increased systems cost. It is therefore desirable to have on-line and continuous access to performance data on influent and effluent quality. Relevant variables to indicate process performance include VFA, COD, alkalinity, sulphate, and, if aerobic post-treatment is considered, total nitrogen, ammonia and nitrate. Recently, mid-IR spectrometry was demonstrated on a pilot scale to be suitable for in-line simultaneous measurement of these variables. This paper describes a full-scale application of the technique to test its ability to monitor continuously and without human intervention the above variables simultaneously in two process streams. For VFA, COD, sulphate, ammonium and TKN good agreement was obtained between in-line and manual measurements. During a period of six months the in-line measurements had to be interrupted several times because of clogging. It appeared that the sample pre-treatment unit was not able to cope with high solids concentrations all the time.
Ialongo, Cristiano; Pieri, Massimo; Bernardini, Sergio
2017-02-01
Saving resources is a paramount issue for the modern laboratory, and new trainable as well as smart technologies can be used to allow the automated instrumentation to manage samples more efficiently in order to achieve streamlined processes. In this regard the serum free light chain (sFLC) testing represents an interesting challenge, as it usually causes using a number of assays before achieving an acceptable result within the analytical range. An artificial neural network based on the multi-layer perceptron (MLP-ANN) was used to infer the starting dilution status of sFLC samples based on the information available through the laboratory information system (LIS). After the learning phase, the MLP-ANN simulation was applied to the nephelometric testing routinely performed in our laboratory on a BN ProSpec® System analyzer (Siemens Helathcare) using the N Latex FLC kit. The MLP-ANN reduced the serum kappa free light chain (κ-FLC) and serum lambda free light chain (λ-FLC) wasted tests by 69.4% and 70.8% with respect to the naïve stepwise dilution scheme used by the automated analyzer, and by 64.9% and 66.9% compared to a "rational" dilution scheme based on a 4-step dilution. Although it was restricted to follow-up samples, the MLP-ANN showed good predictive performance, which alongside the possibility to implement it in any automated system, made it a suitable solution for achieving streamlined laboratory processes and saving resources.
Sample to answer visualization pipeline for low-cost point-of-care blood cell counting
NASA Astrophysics Data System (ADS)
Smith, Suzanne; Naidoo, Thegaran; Davies, Emlyn; Fourie, Louis; Nxumalo, Zandile; Swart, Hein; Marais, Philip; Land, Kevin; Roux, Pieter
2015-03-01
We present a visualization pipeline from sample to answer for point-of-care blood cell counting applications. Effective and low-cost point-of-care medical diagnostic tests provide developing countries and rural communities with accessible healthcare solutions [1], and can be particularly beneficial for blood cell count tests, which are often the starting point in the process of diagnosing a patient [2]. The initial focus of this work is on total white and red blood cell counts, using a microfluidic cartridge [3] for sample processing. Analysis of the processed samples has been implemented by means of two main optical visualization systems developed in-house: 1) a fluidic operation analysis system using high speed video data to determine volumes, mixing efficiency and flow rates, and 2) a microscopy analysis system to investigate homogeneity and concentration of blood cells. Fluidic parameters were derived from the optical flow [4] as well as color-based segmentation of the different fluids using a hue-saturation-value (HSV) color space. Cell count estimates were obtained using automated microscopy analysis and were compared to a widely accepted manual method for cell counting using a hemocytometer [5]. The results using the first iteration microfluidic device [3] showed that the most simple - and thus low-cost - approach for microfluidic component implementation was not adequate as compared to techniques based on manual cell counting principles. An improved microfluidic design has been developed to incorporate enhanced mixing and metering components, which together with this work provides the foundation on which to successfully implement automated, rapid and low-cost blood cell counting tests.
NASA Technical Reports Server (NTRS)
Spivey, Reggie A.; Gilley, Scott; Ostrogorsky, Aleksander; Grugel, Richard; Smith, Guy; Luz, Paul
2003-01-01
The Solidification Using a Baffle in Sealed Ampoules (SUBSA) and Pore Formation and Mobility Investigation (PFMI) furnaces were developed for operation in the International Space Station (ISS) Microgravity Science Glovebox (MSG). Both furnaces were launched to the ISS on STS-111, June 4, 2002, and are currently in use on orbit. The SUBSA furnace provides a maximum temperature of 850 C and can accommodate a metal sample as large as 30 cm long and 12mm in diameter. SUBSA utilizes a gradient freeze process with a minimum cooldown rate of 0.5C per min, and a stability of +/- 0.15C. An 8 cm long transparent gradient zone coupled with a Cohu 3812 camera and quartz ampoule allows for observation and video recording of the solidification process. PFMI is a Bridgman type furnace that operates at a maximum temperature of 130C and can accommodate a sample 23cm long and 10mm in diameter. Two Cohu 3812 cameras mounted 90 deg apart move on a separate translation system which allows for viewing of the sample in the transparent hot zone and gradient zone independent of the furnace translation rate and direction. Translation rates for both the cameras and furnace can be specified from 0.5micrometers/sec to 100 micrometers/sec with a stability of +/-5%. The two furnaces share a Process Control Module (PCM) which controls the furnace hardware, a Data Acquisition Pad (DaqPad) which provides signal condition of thermal couple data, and two Cohu 3812 cameras. The hardware and software allow for real time monitoring and commanding of critical process control parameters. This paper will provide a detailed explanation of the SUBSA and PFMI systems along with performance data and some preliminary results from completed on-orbit processing runs.
Ramakumar, Adarsh; Subramanian, Uma; Prasanna, Pataje G S
2015-11-01
High-throughput individual diagnostic dose assessment is essential for medical management of radiation-exposed subjects after a mass casualty. Cytogenetic assays such as the Dicentric Chromosome Assay (DCA) are recognized as the gold standard by international regulatory authorities. DCA is a multi-step and multi-day bioassay. DCA, as described in the IAEA manual, can be used to assess dose up to 4-6 weeks post-exposure quite accurately but throughput is still a major issue and automation is very essential. The throughput is limited, both in terms of sample preparation as well as analysis of chromosome aberrations. Thus, there is a need to design and develop novel solutions that could utilize extensive laboratory automation for sample preparation, and bioinformatics approaches for chromosome-aberration analysis to overcome throughput issues. We have transitioned the bench-based cytogenetic DCA to a coherent process performing high-throughput automated biodosimetry for individual dose assessment ensuring quality control (QC) and quality assurance (QA) aspects in accordance with international harmonized protocols. A Laboratory Information Management System (LIMS) is designed, implemented and adapted to manage increased sample processing capacity, develop and maintain standard operating procedures (SOP) for robotic instruments, avoid data transcription errors during processing, and automate analysis of chromosome-aberrations using an image analysis platform. Our efforts described in this paper intend to bridge the current technological gaps and enhance the potential application of DCA for a dose-based stratification of subjects following a mass casualty. This paper describes one such potential integrated automated laboratory system and functional evolution of the classical DCA towards increasing critically needed throughput. Published by Elsevier B.V.
Common computational properties found in natural sensory systems
NASA Astrophysics Data System (ADS)
Brooks, Geoffrey
2009-05-01
Throughout the animal kingdom there are many existing sensory systems with capabilities desired by the human designers of new sensory and computational systems. There are a few basic design principles constantly observed among these natural mechano-, chemo-, and photo-sensory systems, principles that have been proven by the test of time. Such principles include non-uniform sampling and processing, topological computing, contrast enhancement by localized signal inhibition, graded localized signal processing, spiked signal transmission, and coarse coding, which is the computational transformation of raw data using broadly overlapping filters. These principles are outlined here with references to natural biological sensory systems as well as successful biomimetic sensory systems exploiting these natural design concepts.
A Synopsis of Technical Issues of Concern for Monitoring Trace Elements in Highway and Urban Runoff
Breault, Robert F.; Granato, Gregory E.
2000-01-01
Trace elements, which are regulated for aquatic life protection, are a primary concern in highway- and urban-runoff studies because stormwater runoff may transport these constituents from the land surface to receiving waters. Many of these trace elements are essential for biological activity and become detrimental only when geologic or anthropogenic sources exceed concentrations beyond ranges typical of the natural environment. The Federal Highway Administration and State Transportation Agencies are concerned about the potential effects of highway runoff on the watershed scale and for the management and protection of watersheds. Transportation agencies need information that is documented as valid, current, and scientifically defensible to support planning and management decisions. There are many technical issues of concern for monitoring trace elements; therefore, trace-element data commonly are considered suspect, and the responsibility to provide data-quality information to support the validity of reported results rests with the data-collection agency. Paved surfaces are fundamentally different physically, hydraulically, and chemically from the natural surfaces typical of most freshwater systems that have been the focus of many traceelement- monitoring studies. Existing scientific conceptions of the behavior of trace elements in the environment are based largely upon research on natural systems, rather than on systems typical of pavement runoff. Additionally, the logistics of stormwater sampling are difficult because of the great uncertainty in the occurrence and magnitude of storm events. Therefore, trace-element monitoring programs may be enhanced if monitoring and sampling programs are automated. Automation would standardize the process and provide a continuous record of the variations in flow and water-quality characteristics. Great care is required to collect and process samples in a manner that will minimize potential contamination or attenuation of trace elements and other sources of bias and variability in the sampling process. Trace elements have both natural and anthropogenic sources that may affect the sampling process, including the sample-collection and handling materials used in many trace-element monitoring studies. Trace elements also react with these materials within the timescales typical for collection, processing and analysis of runoff samples. To study the characteristics and potential effects of trace elements in highway and urban runoff, investigators typically sample one or more operationally defined matrixes including: whole water, dissolved (filtered water), suspended sediment, bottom sediment, biological tissue, and contaminant sources. The sampling and analysis of each of these sample matrixes can provide specific information about the occurrence and distribution of trace elements in runoff and receiving waters. There are, however, technical concerns specific to each matrix that must be understood and addressed through use of proper collection and processing protocols. Valid protocols are designed to minimize inherent problems and to maximize the accuracy, precision, comparability, and representativeness of data collected. Documentation, including information about monitoring protocols, quality assurance and quality control efforts, and ancillary data also is necessary to establish data quality. This documentation is especially important for evaluation of historical traceelement monitoring data, because trace-element monitoring protocols and analysis methods have been constantly changing over the past 30 years.
An introduction to the Marshall information retrieval and display system
NASA Technical Reports Server (NTRS)
1974-01-01
An on-line terminal oriented data storage and retrieval system is presented which allows a user to extract and process information from stored data bases. The use of on-line terminals for extracting and displaying data from the data bases provides a fast and responsive method for obtaining needed information. The system consists of general purpose computer programs that provide the overall capabilities of the total system. The system can process any number of data files via a Dictionary (one for each file) which describes the data format to the system. New files may be added to the system at any time, and reprogramming is not required. Illustrations of the system are shown, and sample inquiries and responses are given.
2018-01-01
ABSTRACT To assess phenotypic bacterial antimicrobial resistance (AMR) in different strata (e.g., host populations, environmental areas, manure, or sewage effluents) for epidemiological purposes, isolates of target bacteria can be obtained from a stratum using various sample types. Also, different sample processing methods can be applied. The MIC of each target antimicrobial drug for each isolate is measured. Statistical equivalence testing of the MIC data for the isolates allows evaluation of whether different sample types or sample processing methods yield equivalent estimates of the bacterial antimicrobial susceptibility in the stratum. We demonstrate this approach on the antimicrobial susceptibility estimates for (i) nontyphoidal Salmonella spp. from ground or trimmed meat versus cecal content samples of cattle in processing plants in 2013-2014 and (ii) nontyphoidal Salmonella spp. from urine, fecal, and blood human samples in 2015 (U.S. National Antimicrobial Resistance Monitoring System data). We found that the sample types for cattle yielded nonequivalent susceptibility estimates for several antimicrobial drug classes and thus may gauge distinct subpopulations of salmonellae. The quinolone and fluoroquinolone susceptibility estimates for nontyphoidal salmonellae from human blood are nonequivalent to those from urine or feces, conjecturally due to the fluoroquinolone (ciprofloxacin) use to treat infections caused by nontyphoidal salmonellae. We also demonstrate statistical equivalence testing for comparing sample processing methods for fecal samples (culturing one versus multiple aliquots per sample) to assess AMR in fecal Escherichia coli. These methods yield equivalent results, except for tetracyclines. Importantly, statistical equivalence testing provides the MIC difference at which the data from two sample types or sample processing methods differ statistically. Data users (e.g., microbiologists and epidemiologists) may then interpret practical relevance of the difference. IMPORTANCE Bacterial antimicrobial resistance (AMR) needs to be assessed in different populations or strata for the purposes of surveillance and determination of the efficacy of interventions to halt AMR dissemination. To assess phenotypic antimicrobial susceptibility, isolates of target bacteria can be obtained from a stratum using different sample types or employing different sample processing methods in the laboratory. The MIC of each target antimicrobial drug for each of the isolates is measured, yielding the MIC distribution across the isolates from each sample type or sample processing method. We describe statistical equivalence testing for the MIC data for evaluating whether two sample types or sample processing methods yield equivalent estimates of the bacterial phenotypic antimicrobial susceptibility in the stratum. This includes estimating the MIC difference at which the data from the two approaches differ statistically. Data users (e.g., microbiologists, epidemiologists, and public health professionals) can then interpret whether that present difference is practically relevant. PMID:29475868
Shakeri, Heman; Volkova, Victoriya; Wen, Xuesong; Deters, Andrea; Cull, Charley; Drouillard, James; Müller, Christian; Moradijamei, Behnaz; Jaberi-Douraki, Majid
2018-05-01
To assess phenotypic bacterial antimicrobial resistance (AMR) in different strata (e.g., host populations, environmental areas, manure, or sewage effluents) for epidemiological purposes, isolates of target bacteria can be obtained from a stratum using various sample types. Also, different sample processing methods can be applied. The MIC of each target antimicrobial drug for each isolate is measured. Statistical equivalence testing of the MIC data for the isolates allows evaluation of whether different sample types or sample processing methods yield equivalent estimates of the bacterial antimicrobial susceptibility in the stratum. We demonstrate this approach on the antimicrobial susceptibility estimates for (i) nontyphoidal Salmonella spp. from ground or trimmed meat versus cecal content samples of cattle in processing plants in 2013-2014 and (ii) nontyphoidal Salmonella spp. from urine, fecal, and blood human samples in 2015 (U.S. National Antimicrobial Resistance Monitoring System data). We found that the sample types for cattle yielded nonequivalent susceptibility estimates for several antimicrobial drug classes and thus may gauge distinct subpopulations of salmonellae. The quinolone and fluoroquinolone susceptibility estimates for nontyphoidal salmonellae from human blood are nonequivalent to those from urine or feces, conjecturally due to the fluoroquinolone (ciprofloxacin) use to treat infections caused by nontyphoidal salmonellae. We also demonstrate statistical equivalence testing for comparing sample processing methods for fecal samples (culturing one versus multiple aliquots per sample) to assess AMR in fecal Escherichia coli These methods yield equivalent results, except for tetracyclines. Importantly, statistical equivalence testing provides the MIC difference at which the data from two sample types or sample processing methods differ statistically. Data users (e.g., microbiologists and epidemiologists) may then interpret practical relevance of the difference. IMPORTANCE Bacterial antimicrobial resistance (AMR) needs to be assessed in different populations or strata for the purposes of surveillance and determination of the efficacy of interventions to halt AMR dissemination. To assess phenotypic antimicrobial susceptibility, isolates of target bacteria can be obtained from a stratum using different sample types or employing different sample processing methods in the laboratory. The MIC of each target antimicrobial drug for each of the isolates is measured, yielding the MIC distribution across the isolates from each sample type or sample processing method. We describe statistical equivalence testing for the MIC data for evaluating whether two sample types or sample processing methods yield equivalent estimates of the bacterial phenotypic antimicrobial susceptibility in the stratum. This includes estimating the MIC difference at which the data from the two approaches differ statistically. Data users (e.g., microbiologists, epidemiologists, and public health professionals) can then interpret whether that present difference is practically relevant. Copyright © 2018 Shakeri et al.
Understanding microbial/DOM interactions using fluorescence and flow cytometry
NASA Astrophysics Data System (ADS)
Fox, Bethany; Rushworth, Cathy; Attridge, John; Anesio, Alexandre; Cox, Tim; Reynolds, Darren
2015-04-01
The transformation and movement of dissolved organic carbon (DOC) within freshwater aquatic systems is an important factor in the global cycling of carbon. DOC within aquatic systems is known to underpin the microbial food web and therefore plays an essential role in supporting and maintaining the aquatic ecosystem. Despite this the interactions between bacteria and dissolved organic matter (DOM) are not well understood, although the literature indicates that the microbial processing of bioavailable DOM is essential during the production of autochthonous, labile, DOM. DOM can be broadly characterised by its fluorescing properties and Coble et al. (2014) define terrestrially derived DOM as exhibiting "peak C" fluorescence, whilst labile microbially derived DOM is defined as showing "peak T" fluorescence. Our work explores the microbial/DOM interactions by analysing aquatic samples using fluorescence excitation and emission matrices (EEMs) in conjunction with microbial consumption of dissolved oxygen. Environmental and synthetic water samples were subjected to fluorescence characterisation using both fluorescence spectroscopy and in situ fluorescence sensors (Chelsea Technologies Group Ltd.). PARAFAC analysis and peak picking were performed on EEMs and compared with flow cytometry data, used to quantify bacterial numbers present within samples. Synthetic samples were created using glucose, glutamic acid, nutrient-rich water and a standard bacterial seed. Synthetic samples were provided with terrestrially derived DOM via the addition of an aliquot of environmental water. Using a closed system approach, samples were incubated over time (up to a maximum of 20 days) and analysed at pre-defined intervals. The main focus of our work is to improve our understanding of microbial/DOM interactions and how these interactions affect both the DOM characteristics and microbial food web in freshwater aquatic systems. The information gained, in relation to the origin, microbial processing and subsequent production of DOM, will inform the development of a new generation of in situ fluorescence sensors. Ultimately, our aim is develop a novel technology that enables the monitoring of ecosystem health in freshwater aquatic systems.
Development of a compact and cost effective multi-input digital signal processing system
NASA Astrophysics Data System (ADS)
Darvish-Molla, Sahar; Chin, Kenrick; Prestwich, William V.; Byun, Soo Hyun
2018-01-01
A prototype digital signal processing system (DSP) was developed using a microcontroller interfaced with a 12-bit sampling ADC, which offers a considerably inexpensive solution for processing multiple detectors with high throughput. After digitization of the incoming pulses, in order to maximize the output counting rate, a simple algorithm was employed for pulse height analysis. Moreover, an algorithm aiming at the real-time pulse pile-up deconvolution was implemented. The system was tested using a NaI(Tl) detector in comparison with a traditional analogue and commercial digital systems for a variety of count rates. The performance of the prototype system was consistently superior to the analogue and the commercial digital systems up to the input count rate of 61 kcps while was slightly inferior to the commercial digital system but still superior to the analogue system in the higher input rates. Considering overall cost, size and flexibility, this custom made multi-input digital signal processing system (MMI-DSP) was the best reliable choice for the purpose of the 2D microdosimetric data collection, or for any measurement in which simultaneous multi-data collection is required.
Laboratory formation of non-cementing, methane hydrate-bearing sands
Waite, William F.; Bratton, Peter M.; Mason, David H.
2011-01-01
Naturally occurring hydrate-bearing sands often behave as though methane hydrate is acting as a load-bearing member of the sediment. Mimicking this behavior in laboratory samples with methane hydrate likely requires forming hydrate from methane dissolved in water. To hasten this formation process, we initially form hydrate in a free-gas-limited system, then form additional hydrate by circulating methane-supersaturated water through the sample. Though the dissolved-phase formation process can theoretically be enhanced by increasing the pore pressure and flow rate and lowering the sample temperature, a more fundamental concern is preventing clogs resulting from inadvertent methane bubble formation in the circulation lines. Clog prevention requires careful temperature control throughout the circulation loop.
Investigation of Zerodur material processing
NASA Technical Reports Server (NTRS)
Johnson, R. Barry
1993-01-01
The Final Report of the Center for Applied Optics (CAO), of The University of Alabama (UAH) study entitled 'Investigation of Zerodur Material Processing' is presented. The objectives of the effort were to prepare glass samples by cutting, grinding, etching, and polishing block Zerodur to desired specifications using equipment located in the optical shop located in the Optical System Branch at NASA/MSFC; characterize samples for subsurface damage and surface roughness; utilize Zerodur samples for coating investigations; and perform investigations into enhanced optical fabrication and metrology techniques. The results of this investigation will be used to support the Advanced X Ray Astrophysics Facility (AXAF) program as well as other NASA/MSFC research programs. The results of the technical effort are presented and discussed.
An Information Storage and Retrieval System for Biological and Geological Data. Interim Report.
ERIC Educational Resources Information Center
Squires, Donald F.
A project is being conducted to test the feasibility of an information storage and retrieval system for museum specimen data, particularly for natural history museums. A pilot data processing system has been developed, with the specimen records from the national collections of birds, marine crustaceans, and rocks used as sample data. The research…
Micro Thermal and Chemical Systems for In Situ Resource Utilization on Mars
NASA Technical Reports Server (NTRS)
Wegeng, Robert S.; Sanders, Gerald
2000-01-01
Robotic sample return missions and postulated human missions to Mars can be greatly aided through the development and utilization of compact chemical processing systems that process atmospheric gases and other indigenous resources to produce hydrocarbon propellants/fuels, oxygen, and other needed chemicals. When used to reduce earth launch mass, substantial cost savings can result. Process Intensification and Process Miniaturization can simultaneously be achieved through the application of microfabricated chemical process systems, based on the rapid heat and mass transport in engineered microchannels. Researchers at NASA's Johnson Space Center (JSC) and the Department of Energy's Pacific Northwest National Laboratory (PNNL) are collaboratively developing micro thermal and chemical systems for NASA's Mission to Mars program. Preliminary results show that many standard chemical process components (e.g., heat exchangers, chemical reactors and chemical separations units) can be reduced in hardware volume without a corresponding reduction in chemical production rates. Low pressure drops are also achievable when appropriate scaling rules are applied. This paper will discuss current progress in the development of engineered microchemical systems for space and terrestrial applications, including fabrication methods, expected operating characteristics, and specific experimental results.
1984-02-15
Directory ... ....... 42 19. Sample Interval Monitor Graph ................. 46 vi vii LIST OF FIGURES P age I. Example of DATA PROFILE Plot...Final Report, AFGL-TR-81-0130, ADA1? 7879 . PART I. RAW DATA TAPE PROCESSING PROCEDURES. 1.1 EXPERIMENT SAMPLING SEQUENCES Due to the changing...Data Quality 79 QQQQ Packed Eltro Data Quality 80 QQQQQQQQ Packed Luxmeter Data Quality 8i QQOQ Packed Night Path Data Quality P? QQQQQQQ Packed Vis
2004-09-01
identification of the lettered features. 2.2 BFIT Sampling Chip The BFIT sampling chip is a flexible patch-like chip with a multilayer polymeric metal...PPy) and glucose oxidase (GOD). The BFIT fabrication process uses SU8 as a principal structural material consisting of five steps (Figure 2). This...process is a subset of an earlier technology developed for the polymer material PDMS.11,12,13,14,15 The first step was the deposition of a Teflon
Enhanced clinical-scale manufacturing of TCR transduced T-cells using closed culture system modules.
Jin, Jianjian; Gkitsas, Nikolaos; Fellowes, Vicki S; Ren, Jiaqiang; Feldman, Steven A; Hinrichs, Christian S; Stroncek, David F; Highfill, Steven L
2018-01-24
Genetic engineering of T-cells to express specific T cell receptors (TCR) has emerged as a novel strategy to treat various malignancies. More widespread utilization of these types of therapies has been somewhat constrained by the lack of closed culture processes capable of expanding sufficient numbers of T-cells for clinical application. Here, we evaluate a process for robust clinical grade manufacturing of TCR gene engineered T-cells. TCRs that target human papillomavirus E6 and E7 were independently tested. A 21 day process was divided into a transduction phase (7 days) and a rapid expansion phase (14 days). This process was evaluated using two healthy donor samples and four samples obtained from patients with epithelial cancers. The process resulted in ~ 2000-fold increase in viable nucleated cells and high transduction efficiencies (64-92%). At the end of culture, functional assays demonstrated that these cells were potent and specific in their ability to kill tumor cells bearing target and secrete large quantities of interferon and tumor necrosis factor. Both phases of culture were contained within closed or semi-closed modules, which include automated density gradient separation and cell culture bags for the first phase and closed GREX culture devices and wash/concentrate systems for the second phase. Large-scale manufacturing using modular systems and semi-automated devices resulted in highly functional clinical-grade TCR transduced T-cells. This process is now in use in actively accruing clinical trials and the NIH Clinical Center and can be utilized at other cell therapy manufacturing sites that wish to scale-up and optimize their processing using closed systems.
Cotton gin drying systems–patterns in fuel energy use
USDA-ARS?s Scientific Manuscript database
Fuel cost increases and consumption variability threaten gin profitability. Twenty-three U.S. cotton gins were audited to elucidate drying system components, layouts and process control strategies that made better use of fuel. Seed cotton samples were obtained before and after each drying system a...
Barnett, J Matthew; Yu, Xiao-Ying; Recknagle, Kurtis P; Glissmeyer, John A
2016-11-01
A planned laboratory space and exhaust system modification to the Pacific Northwest National Laboratory Material Science and Technology Building indicated that a new evaluation of the mixing at the air sampling system location would be required for compliance to ANSI/HPS N13.1-2011. The modified exhaust system would add a third fan, thereby increasing the overall exhaust rate out the stack, thus voiding the previous mixing study. Prior to modifying the radioactive air emissions exhaust system, a three-dimensional computational fluid dynamics computer model was used to evaluate the mixing at the sampling system location. Modeling of the original three-fan system indicated that not all mixing criteria could be met. A second modeling effort was conducted with the addition of an air blender downstream of the confluence of the three fans, which then showed satisfactory mixing results. The final installation included an air blender, and the exhaust system underwent full-scale tests to verify velocity, cyclonic flow, gas, and particulate uniformity. The modeling results and those of the full-scale tests show agreement between each of the evaluated criteria. The use of a computational fluid dynamics code was an effective aid in the design process and allowed the sampling system to remain in its original location while still meeting the requirements for sampling at a well mixed location.
NASA Astrophysics Data System (ADS)
Whitehouse, C. R.; Barnett, S. J.; Soley, D. E. J.; Quarrell, J.; Aldridge, S. J.; Cullis, A. G.; Emeny, M. T.; Johnson, A. D.; Clarke, G. F.; Lamb, W.; Tanner, B. K.; Cottrell, S.; Lunn, B.; Hogg, C.; Hagston, W.
1992-01-01
This paper describes a unique combined UHV MBE growth x-ray topography facility designed to allow the first real-time synchrotron radiation x-ray topography study of strained-layer III-V growth processes. This system will enable unambiguous determination of dislocation nucleation and multiplication processes as a function of controlled variations in growth conditions, and also during post-growth thermal processing. The planned experiments have placed very stringent demands upon the engineering design of the system, and design details regarding the growth chamber; sample manipulator, x-ray optics, and real-time imaging systems are described. Results obtained during a feasibility study are also presented.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Whitehouse, C.R.; Barnett, S.J.; Soley, D.E.J.
1992-01-01
This paper describes a unique combined UHV MBE growth x-ray topography facility designed to allow the first real-time synchrotron radiation x-ray topography study of strained-layer III--V growth processes. This system will enable unambiguous determination of dislocation nucleation and multiplication processes as a function of controlled variations in growth conditions, and also during post-growth thermal processing. The planned experiments have placed very stringent demands upon the engineering design of the system, and design details regarding the growth chamber; sample manipulator, x-ray optics, and real-time imaging systems are described. Results obtained during a feasibility study are also presented.
Greenspoon, S A; Sykes, K L V; Ban, J D; Pollard, A; Baisden, M; Farr, M; Graham, N; Collins, B L; Green, M M; Christenson, C C
2006-12-20
Human genome, pharmaceutical and research laboratories have long enjoyed the application of robotics to performing repetitive laboratory tasks. However, the utilization of robotics in forensic laboratories for processing casework samples is relatively new and poses particular challenges. Since the quantity and quality (a mixture versus a single source sample, the level of degradation, the presence of PCR inhibitors) of the DNA contained within a casework sample is unknown, particular attention must be paid to procedural susceptibility to contamination, as well as DNA yield, especially as it pertains to samples with little biological material. The Virginia Department of Forensic Science (VDFS) has successfully automated forensic casework DNA extraction utilizing the DNA IQ(trade mark) System in conjunction with the Biomek 2000 Automation Workstation. Human DNA quantitation is also performed in a near complete automated fashion utilizing the AluQuant Human DNA Quantitation System and the Biomek 2000 Automation Workstation. Recently, the PCR setup for casework samples has been automated, employing the Biomek 2000 Automation Workstation and Normalization Wizard, Genetic Identity version, which utilizes the quantitation data, imported into the software, to create a customized automated method for DNA dilution, unique to that plate of DNA samples. The PCR Setup software method, used in conjunction with the Normalization Wizard method and written for the Biomek 2000, functions to mix the diluted DNA samples, transfer the PCR master mix, and transfer the diluted DNA samples to PCR amplification tubes. Once the process is complete, the DNA extracts, still on the deck of the robot in PCR amplification strip tubes, are transferred to pre-labeled 1.5 mL tubes for long-term storage using an automated method. The automation of these steps in the process of forensic DNA casework analysis has been accomplished by performing extensive optimization, validation and testing of the software methods.
[Research status and prospects of DNA test on difficult specimens].
Dang, Hua-Wei; Mao, Jiong; Wang, Hui; Huang, Jiang-Ping; Bai, Xiao-Gang
2012-02-01
This paper reviews the advances of DNA detection on three types of difficult biological specimens including degraded samples, trace evidences and mixed samples. The source of different samples, processing methods and announcements were analyzed. New methods such as mitochondrial test system, changing the original experimental conditions, low-volume PCR amplification and new technologies such as whole genome amplification techniques, laser capture micro-dissection, and mini-STR technology in recent years are introduced.
Boe, Kanokwan; Batstone, Damien John; Angelidaki, Irini
2007-03-01
A new method for online measurement of volatile fatty acids (VFA) in anerobic digesters has been developed based on headspace gas chromatography (HSGC). The method applies ex situ VFA stripping with variable headspace volume and gas analysis by gas chromatography-flame ionization detection (GC-FID). In each extraction, digester sample was acidified with H(3)PO(4) and NaHSO(4), then heated to strip the VFA into the gas phase. The gas was sampled in a low friction glass syringe before injected into the GC for measurement. The system has been tested for online monitoring of a lab-scale CSTR reactor treating manure for more than 6 months and has shown good agreement with off-line analysis. The system is capable of measuring individual VFA components. This is of advantage since specific VFA components such as propionic and butyric acid can give extra information about the process status. Another important advantage of this sensor is that there is no filtration, which makes possible application in high solids environments. The system can thus be easily applied in a full-scale biogas reactor by connecting the system to the liquid circulation loop to obtain fresh sample from the reactor. Local calibration is needed but automatic calibration is also possible using standard addition method. Sampling duration is 25-40 min, depending on the washing duration, and sensor response is 10 min. This is appropriate for full-scale reactors, since dynamics within most biogas reactors are of the order of several hours.
A fully automated liquid–liquid extraction system utilizing interface detection
Maslana, Eugene; Schmitt, Robert; Pan, Jeffrey
2000-01-01
The development of the Abbott Liquid-Liquid Extraction Station was a result of the need for an automated system to perform aqueous extraction on large sets of newly synthesized organic compounds used for drug discovery. The system utilizes a cylindrical laboratory robot to shuttle sample vials between two loading racks, two identical extraction stations, and a centrifuge. Extraction is performed by detecting the phase interface (by difference in refractive index) of the moving column of fluid drawn from the bottom of each vial containing a biphasic mixture. The integration of interface detection with fluid extraction maximizes sample throughput. Abbott-developed electronics process the detector signals. Sample mixing is performed by high-speed solvent injection. Centrifuging of the samples reduces interface emulsions. Operating software permits the user to program wash protocols with any one of six solvents per wash cycle with as many cycle repeats as necessary. Station capacity is eighty, 15 ml vials. This system has proven successful with a broad spectrum of both ethyl acetate and methylene chloride based chemistries. The development and characterization of this automated extraction system will be presented. PMID:18924693
NASA Technical Reports Server (NTRS)
Saha, C. P.; Bryson, C. E.; Sarrazin, P.; Blake, D. F.
2005-01-01
Many Mars in situ instruments require fine-grained high-fidelity samples of rocks or soil. Included are instruments for the determination of mineralogy as well as organic and isotopic chemistry. Powder can be obtained as a primary objective of a sample collection system (e.g., by collecting powder as a surface is abraded by a rotary abrasion tool (RAT)), or as a secondary objective (e.g, by collecting drill powder as a core is drilled). In the latter case, a properly designed system could be used to monitor drilling in real time as well as to deliver powder to analytical instruments which would perform complementary analyses to those later performed on the intact core. In addition, once a core or other sample is collected, a system that could transfer intelligently collected subsamples of power from the intact core to a suite of analytical instruments would be highly desirable. We have conceptualized, developed and tested a breadboard Powder Delivery System (PoDS) intended to satisfy the collection, processing and distribution requirements of powder samples for Mars in-situ mineralogic, organic and isotopic measurement instruments.