Science.gov

Sample records for castor end-to-end monitoring

  1. Internet end-to-end performance monitoring for the High Energy Nuclear and Particle Physics community

    SciTech Connect

    Matthews, W.

    2000-02-22

    Modern High Energy Nuclear and Particle Physics (HENP) experiments at Laboratories around the world present a significant challenge to wide area networks. Petabytes (1015) or exabytes (1018) of data will be generated during the lifetime of the experiment. Much of this data will be distributed via the Internet to the experiment's collaborators at Universities and Institutes throughout the world for analysis. In order to assess the feasibility of the computing goals of these and future experiments, the HENP networking community is actively monitoring performance across a large part of the Internet used by its collaborators. Since 1995, the pingER project has been collecting data on ping packet loss and round trip times. In January 2000, there are 28 monitoring sites in 15 countries gathering data on over 2,000 end-to-end pairs. HENP labs such as SLAC, Fermi Lab and CERN are using Advanced Network's Surveyor project and monitoring performance from one-way delay of UDP packets. More recently several HENP sites have become involved with NLANR's active measurement program (AMP). In addition SLAC and CERN are part of the RIPE test-traffic project and SLAC is home for a NIMI machine. The large End-to-end performance monitoring infrastructure allows the HENP networking community to chart long term trends and closely examine short term glitches across a wide range of networks and connections. The different methodologies provide opportunities to compare results based on different protocols and statistical samples. Understanding agreement and discrepancies between results provides particular insight into the nature of the network. This paper will highlight the practical side of monitoring by reviewing the special needs of High Energy Nuclear and Particle Physics experiments and provide an overview of the experience of measuring performance across a large number of interconnected networks throughout the world with various methodologies. In particular, results from each project

  2. Telecommunications end-to-end systems monitoring on TOPEX/Poseidon: Tools and techniques

    NASA Technical Reports Server (NTRS)

    Calanche, Bruno J.

    1994-01-01

    The TOPEX/Poseidon Project Satellite Performance Analysis Team's (SPAT) roles and responsibilities have grown to include functions that are typically performed by other teams on JPL Flight Projects. In particular, SPAT Telecommunication's role has expanded beyond the nominal function of monitoring, assessing, characterizing, and trending the spacecraft (S/C) RF/Telecom subsystem to one of End-to-End Information Systems (EEIS) monitoring. This has been accomplished by taking advantage of the spacecraft and ground data system structures and protocols. By processing both the received spacecraft telemetry minor frame ground generated CRC flags and NASCOM block poly error flags, bit error rates (BER) for each link segment can be determined. This provides the capability to characterize the separate link segments, determine science data recovery, and perform fault/anomaly detection and isolation. By monitoring and managing the links, TOPEX has successfully recovered approximately 99.9 percent of the science data with an integrity (BER) of better than 1 x 10(exp 8). This paper presents the algorithms used to process the above flags and the techniques used for EEIS monitoring.

  3. SME2EM: Smart mobile end-to-end monitoring architecture for life-long diseases.

    PubMed

    Serhani, Mohamed Adel; Menshawy, Mohamed El; Benharref, Abdelghani

    2016-01-01

    Monitoring life-long diseases requires continuous measurements and recording of physical vital signs. Most of these diseases are manifested through unexpected and non-uniform occurrences and behaviors. It is impractical to keep patients in hospitals, health-care institutions, or even at home for long periods of time. Monitoring solutions based on smartphones combined with mobile sensors and wireless communication technologies are a potential candidate to support complete mobility-freedom, not only for patients, but also for physicians. However, existing monitoring architectures based on smartphones and modern communication technologies are not suitable to address some challenging issues, such as intensive and big data, resource constraints, data integration, and context awareness in an integrated framework. This manuscript provides a novel mobile-based end-to-end architecture for live monitoring and visualization of life-long diseases. The proposed architecture provides smartness features to cope with continuous monitoring, data explosion, dynamic adaptation, unlimited mobility, and constrained devices resources. The integration of the architecture׳s components provides information about diseases׳ recurrences as soon as they occur to expedite taking necessary actions, and thus prevent severe consequences. Our architecture system is formally model-checked to automatically verify its correctness against designers׳ desirable properties at design time. Its components are fully implemented as Web services with respect to the SOA architecture to be easy to deploy and integrate, and supported by Cloud infrastructure and services to allow high scalability, availability of processes and data being stored and exchanged. The architecture׳s applicability is evaluated through concrete experimental scenarios on monitoring and visualizing states of epileptic diseases. The obtained theoretical and experimental results are very promising and efficiently satisfy the proposed

  4. Portable air quality sensor unit for participatory monitoring: an end-to-end VESNA-AQ based prototype

    NASA Astrophysics Data System (ADS)

    Vucnik, Matevz; Robinson, Johanna; Smolnikar, Miha; Kocman, David; Horvat, Milena; Mohorcic, Mihael

    2015-04-01

    Key words: portable air quality sensor, CITI-SENSE, participatory monitoring, VESNA-AQ The emergence of low-cost easy to use portable air quality sensors units is opening new possibilities for individuals to assess their exposure to air pollutants at specific place and time, and share this information through the Internet connection. Such portable sensors units are being used in an ongoing citizen science project called CITI-SENSE, which enables citizens to measure and share the data. The project aims through creating citizens observatories' to empower citizens to contribute to and participate in environmental governance, enabling them to support and influence community and societal priorities as well as associated decision making. An air quality measurement system based on VESNA sensor platform was primarily designed within the project for the use as portable sensor unit in selected pilot cities (Belgrade, Ljubljana and Vienna) for monitoring outdoor exposure to pollutants. However, functionally the same unit with different set of sensors could be used for example as an indoor platform. The version designed for the pilot studies was equipped with the following sensors: NO2, O3, CO, temperature, relative humidity, pressure and accelerometer. The personal sensor unit is battery powered and housed in a plastic box. The VESNA-based air quality (AQ) monitoring system comprises the VESNA-AQ portable sensor unit, a smartphone app and the remote server. Personal sensor unit supports wireless connection to an Android smartphone via built-in Wi-Fi. The smartphone in turn serves also as the communication gateway towards the remote server using any of available data connections. Besides the gateway functionality the role of smartphone is to enrich data coming from the personal sensor unit with the GPS location, timestamps and user defined context. This, together with an accelerometer, enables the user to better estimate ones exposure in relation to physical activities, time

  5. Interoperable End-to-End Remote Patient Monitoring Platform based on IEEE 11073 PHD and ZigBee Health Care Profile.

    PubMed

    Clarke, Malcolm; de Folter, Joost; Verma, Vivek; Gokalp, Hulya

    2017-08-07

    This paper describes the implementation of an end-to-end remote monitoring platform based on the IEEE 11073 standards for Personal Health Devices (PHD). It provides an overview of the concepts and approaches and describes how the standard has been optimized for small devices with limited resources of processor, memory and power and that use short range wireless technology. It explains aspects of IEEE 11073, including the Domain Information Model, state model and nomenclature, and how these support its plug-and-play architecture. It shows how these aspects underpin a much larger eco-system of interoperable devices and systems that include IHE PCD-01, HL7 and BlueTooth LE medical devices, and the relationship to the Continua Guidelines, advocating the adoption of data standards and nomenclature to support semantic interoperability between health and ambient assisted living (AAL) in future platforms. The paper further describes the adaptions that have been made in order to implement the standard on the ZigBee Health Care Profile and the experiences of implementing an end-to-end platform that has been deployed to frail elderly patients with chronic disease(s) and patients with diabetes.

  6. End-to-End Commitment

    NASA Technical Reports Server (NTRS)

    Newcomb, John

    2004-01-01

    The end-to-end test would verify the complex sequence of events from lander separation to landing. Due to the large distances involved and the significant delay time in sending a command and receiving verification, the lander needed to operate autonomously after it separated from the orbiter. It had to sense conditions, make decisions, and act accordingly. We were flying into a relatively unknown set of conditions-a Martian atmosphere of unknown pressure, density, and consistency to land on a surface of unknown altitude, and one which had an unknown bearing strength.

  7. TROPOMI end-to-end performance studies

    NASA Astrophysics Data System (ADS)

    Voors, Robert; de Vries, Johan; Veefkind, Pepijn; Gloudemans, Annemieke; Mika, Àgnes; Levelt, Pieternel

    2008-10-01

    The TROPOspheric Monitoring Instrument (TROPOMI) is a UV/VIS/NIR/SWIR non-scanning nadir viewing imaging spectrometer that combines a wide swath (110°) with high spatial resolution (8 x 8 km). Its main heritages are from the Ozone Monitoring Instrument (OMI) and from SCIAMACHY. Since its launch in 2004 OMI has been providing, on a daily basis and on a global scale, a wealth of data on ozone, NO2 and minor trace gases, aerosols and local pollution, a scanning spectrometer launched in 2004. The TROPOMI UV/VIS/NIR and SWIR heritage is a combination of OMI and SCIAMACHY. In the framework of development programs for a follow-up mission for the successful Ozone Monitoring Instrument, we have developed the so-called TROPOMI Integrated Development Environment. This is a GRID based software simulation tool for OMI follow-up missions. It includes scene generation, an instrument simulator, a level 0-1b processing chain, as well as several level 1b-2 processing chains. In addition it contains an error-analyzer, i.e. a tool to feedback the level 2 results to the input of the scene generator. The paper gives a description of the TROPOMI instrument and focuses on design aspects as well as on the performance, as tested in the end-to-end development environment TIDE.

  8. TOWARD END-TO-END MODELING FOR NUCLEAR EXPLOSION MONITORING: SIMULATION OF UNDERGROUND NUCLEAR EXPLOSIONS AND EARTHQUAKES USING HYDRODYNAMIC AND ANELASTIC SIMULATIONS, HIGH-PERFORMANCE COMPUTING AND THREE-DIMENSIONAL EARTH MODELS

    SciTech Connect

    Rodgers, A; Vorobiev, O; Petersson, A; Sjogreen, B

    2009-07-06

    This paper describes new research being performed to improve understanding of seismic waves generated by underground nuclear explosions (UNE) by using full waveform simulation, high-performance computing and three-dimensional (3D) earth models. The goal of this effort is to develop an end-to-end modeling capability to cover the range of wave propagation required for nuclear explosion monitoring (NEM) from the buried nuclear device to the seismic sensor. The goal of this work is to improve understanding of the physical basis and prediction capabilities of seismic observables for NEM including source and path-propagation effects. We are pursuing research along three main thrusts. Firstly, we are modeling the non-linear hydrodynamic response of geologic materials to underground explosions in order to better understand how source emplacement conditions impact the seismic waves that emerge from the source region and are ultimately observed hundreds or thousands of kilometers away. Empirical evidence shows that the amplitudes and frequency content of seismic waves at all distances are strongly impacted by the physical properties of the source region (e.g. density, strength, porosity). To model the near-source shock-wave motions of an UNE, we use GEODYN, an Eulerian Godunov (finite volume) code incorporating thermodynamically consistent non-linear constitutive relations, including cavity formation, yielding, porous compaction, tensile failure, bulking and damage. In order to propagate motions to seismic distances we are developing a one-way coupling method to pass motions to WPP (a Cartesian anelastic finite difference code). Preliminary investigations of UNE's in canonical materials (granite, tuff and alluvium) confirm that emplacement conditions have a strong effect on seismic amplitudes and the generation of shear waves. Specifically, we find that motions from an explosion in high-strength, low-porosity granite have high compressional wave amplitudes and weak shear

  9. Castor

    USDA-ARS?s Scientific Manuscript database

    Castor (Ricinus communis L.) has the potential to become the premier oil crop for industrial oil production across the glove and is an ideal candidate for production of high value, industrial oil because of the very high oil content (48-60%) of the seed, the extremely high levels of potential oil pr...

  10. Applying Trustworthy Computing to End-to-End Electronic Voting

    ERIC Educational Resources Information Center

    Fink, Russell A.

    2010-01-01

    "End-to-End (E2E)" voting systems provide cryptographic proof that the voter's intention is captured, cast, and tallied correctly. While E2E systems guarantee integrity independent of software, most E2E systems rely on software to provide confidentiality, availability, authentication, and access control; thus, end-to-end integrity is not…

  11. Applying Trustworthy Computing to End-to-End Electronic Voting

    ERIC Educational Resources Information Center

    Fink, Russell A.

    2010-01-01

    "End-to-End (E2E)" voting systems provide cryptographic proof that the voter's intention is captured, cast, and tallied correctly. While E2E systems guarantee integrity independent of software, most E2E systems rely on software to provide confidentiality, availability, authentication, and access control; thus, end-to-end integrity is not…

  12. End-to-end attraction of duplex DNA

    PubMed Central

    Maffeo, Christopher; Luan, Binquan; Aksimentiev, Aleksei

    2012-01-01

    Recent experiments [Nakata, M. et al., End-to-end stacking and liquid crystal condensation of 6 to 20 basepair DNA duplexes. Science 2007; 318:1276–1279] have demonstrated spontaneous end-to-end association of short duplex DNA fragments into long rod-like structures. By means of extensive all-atom molecular dynamic simulations, we characterized end-to-end interactions of duplex DNA, quantitatively describing the forces, free energy and kinetics of the end-to-end association process. We found short DNA duplexes to spontaneously aggregate end-to-end when axially aligned in a small volume of monovalent electrolyte. It was observed that electrostatic repulsion of 5′-phosphoryl groups promoted the formation of aggregates in a conformation similar to the B-form DNA double helix. Application of an external force revealed that rupture of the end-to-end assembly occurs by the shearing of the terminal base pairs. The standard binding free energy and the kinetic rates of end-to-end association and dissociation processes were estimated using two complementary methods: umbrella sampling simulations of two DNA fragments and direct observation of the aggregation process in a system containing 458 DNA fragments. We found the end-to-end force to be short range, attractive, hydrophobic and only weakly dependent on the ion concentration. The relation between the stacking free energy and end-to-end attraction is discussed as well as possible roles of the end-to-end interaction in biological and nanotechnological systems. PMID:22241779

  13. Standardizing an End-to-end Accounting Service

    NASA Technical Reports Server (NTRS)

    Greenberg, Edward; Kazz, Greg

    2006-01-01

    Currently there are no space system standards available for space agencies to accomplish end-to-end accounting. Such a standard does not exist for spacecraft operations nor for tracing the relationship between the mission planning activities, the command sequences designed to perform those activities, the commands formulated to initiate those activities and the mission data and specifically the mission data products created by those activities. In order for space agencies to cross-support one another for data accountability/data tracing and for inter agency spacecraft to interoperate with each other, an international CCSDS standard for end-to-end data accountability/tracing needs to be developed. We will first describe the end-to-end accounting service model and functionality that supports the service. This model will describe how science plans that are ultimately transformed into commands can be associated with the telemetry products generated as a result of their execution. Moreover, the interaction between end-to-end accounting and service management will be explored. Finally, we will show how the standard end-to-end accounting service can be applied to a real life flight project i.e., the Mars Reconnaissance Orbiter project.

  14. Standardizing an End-to-end Accounting Service

    NASA Technical Reports Server (NTRS)

    Greenberg, Edward; Kazz, Greg

    2006-01-01

    Currently there are no space system standards available for space agencies to accomplish end-to-end accounting. Such a standard does not exist for spacecraft operations nor for tracing the relationship between the mission planning activities, the command sequences designed to perform those activities, the commands formulated to initiate those activities and the mission data and specifically the mission data products created by those activities. In order for space agencies to cross-support one another for data accountability/data tracing and for inter agency spacecraft to interoperate with each other, an international CCSDS standard for end-to-end data accountability/tracing needs to be developed. We will first describe the end-to-end accounting service model and functionality that supports the service. This model will describe how science plans that are ultimately transformed into commands can be associated with the telemetry products generated as a result of their execution. Moreover, the interaction between end-to-end accounting and service management will be explored. Finally, we will show how the standard end-to-end accounting service can be applied to a real life flight project i.e., the Mars Reconnaissance Orbiter project.

  15. Left circumflex artery aneurysm: the end-to-end reconstruction.

    PubMed

    Cuttone, Fabio; Guilbeau-Frugier, Céline; Roncalli, Jérome; Glock, Yves

    2015-08-01

    This report describes a surgical myocardial revascularization procedure for a huge, atherosclerotic left circumflex coronary artery aneurysm. The technique proposed in this paper is based on the isolation of the aneurysm followed by the aneurysmectomy and a complete reconstruction of the circumflex artery by an end-to-end anastomosis.

  16. End-to-end azido-pinned interlocking lanthanide squares.

    PubMed

    Li, Xiao-Lei; Wu, Jianfeng; Zhao, Lang; Shi, Wei; Cheng, Peng; Tang, Jinkui

    2017-03-09

    A rare end-to-end azido-pinned interlocking lanthanide square was self-assembled using a ditopic Schiff-base (H2L) and NaN3 as ligands. Obvious ferromagnetic interaction and a record anisotropy barrier of 152(4) K among lanthanide azido-bridged SMMs in a zero dc field were observed.

  17. Combining Simulation Tools for End-to-End Trajectory Optimization

    NASA Technical Reports Server (NTRS)

    Whitley, Ryan; Gutkowski, Jeffrey; Craig, Scott; Dawn, Tim; Williams, Jacobs; Stein, William B.; Litton, Daniel; Lugo, Rafael; Qu, Min

    2015-01-01

    Trajectory simulations with advanced optimization algorithms are invaluable tools in the process of designing spacecraft. Due to the need for complex models, simulations are often highly tailored to the needs of the particular program or mission. NASA's Orion and SLS programs are no exception. While independent analyses are valuable to assess individual spacecraft capabilities, a complete end-to-end trajectory from launch to splashdown maximizes potential performance and ensures a continuous solution. In order to obtain end-to-end capability, Orion's in-space tool (Copernicus) was made to interface directly with the SLS's ascent tool (POST2) and a new tool to optimize the full problem by operating both simulations simultaneously was born.

  18. End-to-end simulations for the LISA Technology Package

    NASA Astrophysics Data System (ADS)

    Hannen, V. M.; Smit, M.; Hoyng, P.; Selig, A.; Schleicher, A.

    2003-05-01

    We present an end-to-end simulation facility which has been developed in the framework of the LISA Technology Package (LTP) architect study for SMART-2, the technology demonstration mission that precedes LISA. The simulator evolves positions and orientations of the spacecraft and two test masses contained in the inertial sensors of LTP under the influence of external and internal forces and torques and under the influence of control loops for satellite drag-free control (DFC) and electrostatic test mass suspension. Altogether, a coupled system with 18 degrees of freedom is solved numerically. A series of test runs has been performed to verify the correct functioning of the various models contained in the end-to-end simulator and to provide a preliminary assessment of the performance of DFC algorithms and control laws for test mass suspension, which are currently foreseen for use in the basic operation mode of LTP.

  19. LWS/SET End-to-End Data System

    NASA Technical Reports Server (NTRS)

    Giffin, Geoff; Sherman, Barry; Colon, Gilberto (Technical Monitor)

    2002-01-01

    This paper describes the concept for the End-to-End Data System that will support NASA's Living With a Star Space Environment Testbed missions. NASA has initiated the Living With a Star (LWS) Program to develop a better scientific understanding to address the aspects of the connected Sun-Earth system that affect life and society. A principal goal of the program is to bridge the gap.between science, engineering, and user application communities. The Space Environment Testbed (SET) Project is one element of LWS. The Project will enable future science, operational, and commercial objectives in space and atmospheric environments by improving engineering approaches to the accommodation and/or mitigation of the effects of solar variability on technological systems. The End-to-end data system allows investigators to access the SET control center, command their experiments, and receive data from their experiments back at their home facility, using the Internet. The logical functioning of major components of the end-to-end data system are described, including the GSFC Payload Operations Control Center (POCC), SET Payloads, the GSFC SET Simulation Lab, SET Experiment PI Facilities, and Host Systems. Host Spacecraft Operations Control Centers (SOCC) and the Host Spacecraft are essential links in the end-to-end data system, but are not directly under the control of the SET Project. Formal interfaces will be established between these entities and elements of the SET Project. The paper describes data flow through the system, from PI facilities connecting to the SET operations center via the Internet, communications to SET carriers and experiments via host systems, to telemetry returns to investigators from their flight experiments. It also outlines the techniques that will be used to meet mission requirements, while holding development and operational costs to a minimum. Additional information is included in the original extended abstract.

  20. Measurements and analysis of end-to-end Internet dynamics

    SciTech Connect

    Paxson, Vern

    1997-04-01

    Accurately characterizing end-to-end Internet dynamics - the performance that a user actually obtains from the lengthy series of network links that comprise a path through the Internet - is exceptionally difficult, due to the network`s immense heterogeneity. At the heart of this work is a `measurement framework` in which a number of sites around the Internet host a specialized measurement service. By coordinating `probes` between pairs of these sites one can measure end-to-end behavior along O(N2) paths for a framework consisting of N sites. Consequently, one obtains a superlinear scaling that allows measuring a rich cross-section of Internet behavior without requiring huge numbers of observation points. 37 sites participated in this study, allowing the author to measure more than 1,000 distinct Internet paths. The first part of this work looks at the behavior of end-to-end routing: the series of routers over which a connection`s packets travel. Based on 40,000 measurements made using this framework, the author analyzes: routing `pathologies` such as loops, outages, and flutter; the stability of routes over time; and the symmetry of routing along the two directions of an end-to-end path. The author finds that pathologies increased significantly over the course of 1995 and that Internet paths are heavily dominated by a single route. The second part of this work studies end-to-end Internet packet dynamics. The author analyzes 20,000 TCP transfers of 100 Kbyte each to investigate the performance of both the TCP endpoints and the Internet paths. The measurements used for this part of the study are much richer than those for the first part, but require a great degree of attention to issues of calibration, which are addressed by applying self-consistency checks to the measurements whenever possible. The author finds that packet filters are capable of a wide range of measurement errors, some of which, if undetected, can significantly taint subsequent analysis.

  1. OGC standards for end-to-end sensor network integration

    NASA Astrophysics Data System (ADS)

    Headley, K. L.; Broering, A.; O'Reilly, T. C.; Toma, D.; Del Rio, J.; Bermudez, L. E.; Zedlitz, J.; Johnson, G.; Edgington, D.

    2010-12-01

    technology, and can communicate with any sensor whose protocol can be described by a SID. The SID interpreter transfers retrieved sensor data to a Sensor Observation Service, and transforms tasks submitted to a Sensor Planning Service to actual sensor commands. The proposed SWE PUCK protocol complements SID by providing a standard way to associate a sensor with a SID, thereby completely automating the sensor integration process. PUCK protocol is implemented in sensor firmware, and provides a means to retrieve a universally unique identifer, metadata and other information from the device itself through its communication interface. Thus the SID interpreter can retrieve a SID directly from the sensor through PUCK protocol. Alternatively the interpreter can retrieve the sensor’s SID from an external source, based on the unique sensor ID provided by PUCK protocol. In this presentation, we describe the end-to-end integration of several commercial oceanographic instruments into a sensor network using PUCK, SID and SWE services. We also present a user-friendly, graphical tool to generate SIDs and tools to visualize sensor data.

  2. End-to-end network/application performance troubleshooting methodology

    SciTech Connect

    Wu, Wenji; Bobyshev, Andrey; Bowden, Mark; Crawford, Matt; Demar, Phil; Grigaliunas, Vyto; Grigoriev, Maxim; Petravick, Don; /Fermilab

    2007-09-01

    The computing models for HEP experiments are globally distributed and grid-based. Obstacles to good network performance arise from many causes and can be a major impediment to the success of the computing models for HEP experiments. Factors that affect overall network/application performance exist on the hosts themselves (application software, operating system, hardware), in the local area networks that support the end systems, and within the wide area networks. Since the computer and network systems are globally distributed, it can be very difficult to locate and identify the factors that are hurting application performance. In this paper, we present an end-to-end network/application performance troubleshooting methodology developed and in use at Fermilab. The core of our approach is to narrow down the problem scope with a divide and conquer strategy. The overall complex problem is split into two distinct sub-problems: host diagnosis and tuning, and network path analysis. After satisfactorily evaluating, and if necessary resolving, each sub-problem, we conduct end-to-end performance analysis and diagnosis. The paper will discuss tools we use as part of the methodology. The long term objective of the effort is to enable site administrators and end users to conduct much of the troubleshooting themselves, before (or instead of) calling upon network and operating system 'wizards,' who are always in short supply.

  3. Constructing end-to-end models using ECOPATH data

    NASA Astrophysics Data System (ADS)

    Steele, John H.; Ruzicka, James J.

    2011-09-01

    The wide availability of ECOPATH data sets provides a valuable resource for the comparative analysis of marine ecosystems. We show how to derive a bottom-up transform from the top-down ECOPATH; couple this to a simple microbial web with physical forcing; and use the end-to-end model (E2E) for scenario construction. This steady state format also provides a framework and initial conditions for different dynamic simulations. This model can be applied to shelf ecosystems with a wide range of physical forcing, coupled benthic/pelagic food webs, and nutrient recycling. We illustrate the general application and the specific problems by transforming an ECOPATH model for the Northern California Current (NCC). We adapt results on the upwelling regime to provide estimates of physical fluxes and use these to show the consequences of different upwelling rates combined with variable retention mechanism for plankton, for the productivity of fish and other top predators; and for the resilience of the ecosystem. Finally we show how the effects of inter-annual to decadal variations in upwelling on fishery yields can be studied using dynamic simulations with different prey-predator relations. The general conclusion is that the nature of the physical regimes for shelf ecosystems cannot be ignored in comparing end-to-end representations of these food webs.

  4. On routing algorithms with end-to-end delay guarantees

    SciTech Connect

    Rao, N.S.V.; Batsell, S.G.

    1998-11-01

    The authors consider the transmission of a message of size r from a source to a destination with guarantees on the end-to-end delay over a computer network with n nodes and m links. There are three sources of delays: (a) propagation delays along the links, (b) delays due to bandwidth availability on the links, and (c) queuing delays at the intermediate nodes. First, the authors consider that delays on various links and nodes are given as functions of the message size. If the delay in (b) is a non-increasing function of the bandwidth, they propose O(m{sup 2} + mn log n) time algorithm to compute a path with the minimum end-to-end delay for any given message size r. They then consider that the queuing delay in (c) is a random variable correlated with the message size according to an unknown distribution. At each node, the measurements of queuing delays and message sizes are available. They propose two algorithms to compute paths whose delays are close to optimal delays with a high probability, irrespective of the distribution of the delays, and based entirely on the measurements of sufficient size.

  5. Cyberinfrastructure for End-to-End Environmental Explorations

    NASA Astrophysics Data System (ADS)

    Merwade, V.; Kumar, S.; Song, C.; Zhao, L.; Govindaraju, R.; Niyogi, D.

    2007-12-01

    The design and implementation of a cyberinfrastructure for End-to-End Environmental Exploration (C4E4) is presented. The C4E4 framework addresses the need for an integrated data/computation platform for studying broad environmental impacts by combining heterogeneous data resources with state-of-the-art modeling and visualization tools. With Purdue being a TeraGrid Resource Provider, C4E4 builds on top of the Purdue TeraGrid data management system and Grid resources, and integrates them through a service-oriented workflow system. It allows researchers to construct environmental workflows for data discovery, access, transformation, modeling, and visualization. Using the C4E4 framework, we have implemented an end-to-end SWAT simulation and analysis workflow that connects our TeraGrid data and computation resources. It enables researchers to conduct comprehensive studies on the impact of land management practices in the St. Joseph watershed using data from various sources in hydrologic, atmospheric, agricultural, and other related disciplines.

  6. The CarbonSat End-to-End Simulator

    NASA Astrophysics Data System (ADS)

    Bramstedt, Klaus; Noel, Stefan; Bovensmann, Heinrich; Reuter, Max; Burrows, John P.; Jurado Lozano, Pedro Jose; Meijer, Yasjka; Loescher, Armin; Acarreta, Juan R.; Sturm, Philipp; Tesmer, Volker; Sanchez Monero, Ana Maria; Atapuerca Rodreiguez de Dios, Francisco Javier; Toledano Sanchez, Daniel; Boesch, Hartmut

    2016-08-01

    The objective of the CarbonSat mission is to improve our knowledge on natural and anthropogenic sources and sinks of CO2 and CH4. CarbonSat was one of the two candidate missions selected for definition studies for becoming Earth Explorer 8 (EE8).The CarbonSat End-to-End Simulator (CSE2ES) simulates the full data flow of the mission with a set of modules embedded in ESA's generic simulation framework OpenSF. A Geometry Module (GM) defines the orbital geometry and related parameters. A Scene Generation Module (SGM) provides simulated radiances and irradiances for the selected scenes. The Level 1 Module (L1M) compromises the instrument simulator and the Level 1b processor, and provide as main output calibrated spectra. The L1M is implemented in two versions, reflecting the instrument concepts from the two competing industrial system studies. The Level 2 Retrieval Module (L2M) performs the retrieval from the input level 1b spectra to the atmospheric parameters (CO2 and CH4).In this paper, we show sensitivity studies with respect to atmospheric parameters, simulations along the orbit and a case study for the detection of a point source emitting carbon dioxide. In summary, the end-to-end simulation with CSE2ES proves the capability of the CarbonSat concept to reach its requirements.

  7. Miniature modular microwave end-to-end receiver

    NASA Technical Reports Server (NTRS)

    Sukamto, Lin M. (Inventor); Cooley, Thomas W. (Inventor); Janssen, Michael A. (Inventor); Parks, Gary S. (Inventor)

    1993-01-01

    An end-to-end microwave receiver system contained in a single miniature hybrid package mounted on a single heatsink is presented. It includes an input end connected to a microwave receiver antenna and an output end which produces a digital count proportional to the amplitude of a signal of a selected microwave frequency band received at the antenna and corresponding to one of the water vapor absorption lines near frequencies of 20 GHz or 30 GHz. The hybrid package is on the order of several centimeters in length and a few centimeters in height and width. The package includes an L-shaped carrier having a base surface, a vertical wall extending up from the base surface and forming a corner therewith, and connection pins extending through the vertical wall. Modular blocks rest on the base surface against the vertical wall and support microwave monolithic integrated circuits on top surfaces thereof connected to the external connection pins. The modular blocks lie end-to-end on the base surface so as to be modularly removable by sliding along the base surface beneath the external connection pins away from the vertical wall.

  8. Toward End-to-End Face Recognition Through Alignment Learning

    NASA Astrophysics Data System (ADS)

    Zhong, Yuanyi; Chen, Jiansheng; Huang, Bo

    2017-08-01

    Plenty of effective methods have been proposed for face recognition during the past decade. Although these methods differ essentially in many aspects, a common practice of them is to specifically align the facial area based on the prior knowledge of human face structure before feature extraction. In most systems, the face alignment module is implemented independently. This has actually caused difficulties in the designing and training of end-to-end face recognition models. In this paper we study the possibility of alignment learning in end-to-end face recognition, in which neither prior knowledge on facial landmarks nor artificially defined geometric transformations are required. Specifically, spatial transformer layers are inserted in front of the feature extraction layers in a Convolutional Neural Network (CNN) for face recognition. Only human identity clues are used for driving the neural network to automatically learn the most suitable geometric transformation and the most appropriate facial area for the recognition task. To ensure reproducibility, our model is trained purely on the publicly available CASIA-WebFace dataset, and is tested on the Labeled Face in the Wild (LFW) dataset. We have achieved a verification accuracy of 99.08\\% which is comparable to state-of-the-art single model based methods.

  9. Euclid end-to-end straylight performance assessment

    NASA Astrophysics Data System (ADS)

    Gaspar Venancio, Luis M.; Pachot, Charlotte; Carminati, Lionel; Lorenzo Alvarez, Jose; Amiaux, Jérôme; Prieto, Eric; Bonino, Luciana; Salvignol, Jean-Christophe; Short, Alex; Boenke, Tobias; Strada, Paulo; Laureijs, Rene

    2016-07-01

    In the Euclid mission the straylight has been identified at an early stage as the main driver for the final imaging quality of the telescope. The assessment by simulation of the final straylight in the focal plane of both instruments in Euclid's payload have required a complex workflow involving all stakeholders in the mission, from industry to the scientific community. The straylight is defined as a Normalized Detector Irradiance (NDI) which is a convenient definition tool to separate the contributions of the telescope and of the instruments. The end-to-end straylight of the payload is then simply the sum of the NDIs of the telescope and of each instrument. The NDIs for both instruments are presented in this paper for photometry and spectrometry.

  10. An end to end secure CBIR over encrypted medical database.

    PubMed

    Bellafqira, Reda; Coatrieux, Gouenou; Bouslimi, Dalel; Quellec, Gwenole

    2016-08-01

    In this paper, we propose a new secure content based image retrieval (SCBIR) system adapted to the cloud framework. This solution allows a physician to retrieve images of similar content within an outsourced and encrypted image database, without decrypting them. Contrarily to actual CBIR approaches in the encrypted domain, the originality of the proposed scheme stands on the fact that the features extracted from the encrypted images are themselves encrypted. This is achieved by means of homomorphic encryption and two non-colluding servers, we however both consider as honest but curious. In that way an end to end secure CBIR process is ensured. Experimental results carried out on a diabetic retinopathy database encrypted with the Paillier cryptosystem indicate that our SCBIR achieves retrieval performance as good as if images were processed in their non-encrypted form.

  11. Response to MRO's end-to-end data accountability challenges

    NASA Technical Reports Server (NTRS)

    Lee, Young H.

    2005-01-01

    (MRO) on August 12, 2005. It carries six science instruments and three engineering payloads. Because MRO will produce an unprecedented number of science products, it will transmit a much higher data volume via high data rate than any other deep space mission to date. Keeping track of MRO products as well as relay products would be a daunting, expensive task without a well-planned data-product tracking strategy. To respond to this challenge, the MRO project developed the End-to- End Data Accountability System by utilizing existing information available from both ground and flight elements. Therefore, a capability to perform first-order problem diagnosis is essential in order for MRO to answer the questions, where is my data? and when will my data be available? This paper details the approaches taken, design and implementation of the tools, procedures and teams that track data products from the time they are predicted until they arrive in the hands of the end users.

  12. Recirculating Linac Acceleration - End-to-End Simulation

    SciTech Connect

    Alex Bogacz

    2010-03-01

    A conceptual design of a high-pass-number Recirculating Linear Accelerator (RLA) for muons is presented. The scheme involves three superconducting linacs (201 MHz): a single pass linear Pre-accelerator followed by a pair multi-pass (4.5-pass) 'Dogbone' RLAs. Acceleration starts after ionization cooling at 220 MeV/c and proceeds to 12.6 GeV. The Pre-accelerator captures a large muon phase space and accelerates muons to relativistic energies, while adiabatically decreasing the phase-space volume, so that effective acceleration in the RLA is possible. The RLA further compresses and shapes up the longitudinal and transverse phase-spaces, while increasing the energy. Appropriate choice of multi-pass linac optics based on FODO focusing assures large number of passes in the RLA. The proposed 'Dogbone' configuration facilitates simultaneous acceleration of both mu± species through the requirement of mirror symmetric optics of the return 'droplet' arcs. Finally, presented end-to-end simulation validates the efficiency and acceptance of the accelerator system.

  13. Key management for large scale end-to-end encryption

    SciTech Connect

    Witzke, E.L.

    1994-07-01

    Symmetric end-to-end encryption requires separate keys for each pair of communicating confidants. This is a problem of Order N{sup 2}. Other factors, such as multiple sessions per pair of confidants and multiple encryption points in the ISO Reference Model complicate key management by linear factors. Public-key encryption can reduce the number of keys managed to a linear problem which is good for scaleability of key management, but comes with complicating issues and performance penalties. Authenticity is the primary ingredient of key management. If each potential pair of communicating confidants can authenticate data from each other, then any number of public encryption keys of any type can be communicated with requisite integrity. These public encryption keys can be used with the corresponding private keys to exchange symmetric cryptovariables for high data rate privacy protection. The Digital Signature Standard (DSS), which has been adopted by the United States Government, has both public and private components, similar to a public-key cryptosystem. The Digital Signature Algorithm of the DSS is intended for authenticity but not for secrecy. In this paper, the authors will show how the use of the Digital Signature Algorithm combined with both symmetric and asymmetric (public-key) encryption techniques can provide a practical solution to key management scaleability problems, by reducing the key management complexity to a problem of order N, without sacrificing the encryption speed necessary to operate in high performance networks.

  14. End-to-end Adversarial Retinal Image Synthesis.

    PubMed

    Costa, Pedro; Galdran, Adrian; Meyer, Maria Ines; Niemeijer, Meindert; Abramoff, Michael; Mendonca, Ana Maria; Campilho, Aurelio

    2017-10-02

    In medical image analysis applications, the availability of large amounts of annotated data is becoming increasingly critical. However, annotated medical data is often scarce and costly to obtain. In this paper, we address the problem of synthesizing retinal color images by applying recent techniques based on adversarial learning. In this setting, a generative model is trained to maximize a loss function provided by a second model attempting to classify its output into real or synthetic. In particular, we propose to implement an adversarial autoencoder for the task of retinal vessel network synthesis. We use the generated vessel trees as an intermediate stage for the generation of color retinal images, which is accomplished with a Generative Adversarial Network. Both models require the optimization of almost everywhere differentiable loss functions, which allows us to train them jointly. The resulting model offers an end-to-end retinal image synthesis system capable of generating as many retinal images as the user requires, with their corresponding vessel networks, by sampling from a simple probability distribution that we impose to the associated latent space. We show that the learned latent space contains a well-defined semantic structure, implying that we can perform calculations in the space of retinal images, e.g., smoothly interpolating new data points between two retinal images. Visual and quantitative results demonstrate that the synthesized images are substantially different from those in the training set, while being also anatomically consistent and displaying a reasonable visual quality.

  15. Integration proposal through standard-based design of an end-to-end platform for p-Health environments.

    PubMed

    Martíínez, I; Trigo, J D; Martínez-Espronceda, M; Escayola, J; Muñoz, P; Serrano, L; García, J

    2009-01-01

    Interoperability among medical devices and compute engines in the personal environment of the patient, and with healthcare information systems in the remote monitoring and management process is a key need that requires developments supported on standard-based design. Even though there have been some international initiatives to combine different standards, the vision of an entire end-to-end standard-based system is the next challenge. This paper presents the implementation guidelines of a ubiquitous platform for Personal Health (p-Health). It is standard-based using the two main medical norms in this context: ISO/IEEE11073 in the patient environment for medical device interoperability, and EN13606 to allow the interoperable communication of the Electronic Healthcare Record of the patient. Furthermore, the proposal of a new protocol for End-to-End Standard Harmonization (E2ESHP) is presented in order to make possible the end-to-end standard integration. The platform has been designed to comply with the last ISO/IEEE11073 and EN13606 available versions, and tested in a laboratory environment as a proof-of-concept to illustrate its feasibility as an end-to-end standard-based solution.

  16. Experimental demonstration of software defined data center optical networks with Tbps end-to-end tunability

    NASA Astrophysics Data System (ADS)

    Zhao, Yongli; Zhang, Jie; Ji, Yuefeng; Li, Hui; Wang, Huitao; Ge, Chao

    2015-10-01

    The end-to-end tunability is important to provision elastic channel for the burst traffic of data center optical networks. Then, how to complete the end-to-end tunability based on elastic optical networks? Software defined networking (SDN) based end-to-end tunability solution is proposed for software defined data center optical networks, and the protocol extension and implementation procedure are designed accordingly. For the first time, the flexible grid all optical networks with Tbps end-to-end tunable transport and switch system have been online demonstrated for data center interconnection, which are controlled by OpenDayLight (ODL) based controller. The performance of the end-to-end tunable transport and switch system has been evaluated with wavelength number tuning, bit rate tuning, and transmit power tuning procedure.

  17. Understanding TCP over TCP: effects of TCP tunneling on end-to-end throughput and latency

    NASA Astrophysics Data System (ADS)

    Honda, Osamu; Ohsaki, Hiroyuki; Imase, Makoto; Ishizuka, Mika; Murayama, Junichi

    2005-10-01

    TCP tunnel is a technology that aggregates and transfers packets sent between end hosts as a single TCP connection. By using a TCP tunnel, the fairness among aggregated flows can be improved and several protocols can be transparently transmitted through a firewall. Currently, many applications such as SSH, VTun, and HTun use a TCP tunnel. However, since most applications running on end hosts generally use TCP, two TCP congestion controls (i.e., end-to-end TCP and tunnel TCP) operate simultaneously and interfere each other. Under certain conditions, it has been known that using a TCP tunnel severely degrades the end-to-end TCP performance. Namely, it has known that using a TCP tunnel drastically degrades the end-to-end TCP throughput for some time, which is called TCP meltdown problem. On the contrary, under other conditions, it has been known that using a TCP tunnel significantly improves the end-to-end TCP performance. However, it is still an open issue --- how, when, and why is a TCP tunnel malicious for end-to-end TCP performance? In this paper, we therefore investigate effect of TCP tunnel on end-to-end TCP performance using simulation experiments. Specifically, we quantitatively reveal effects of several factors (e.g., the propagation delay, usage of SACK option, TCP socket buffer size, and sender buffer size of TCP tunnel) on performance of end-to-end TCP and tunnel TCP.

  18. A Workflow-based Intelligent Network Data Movement Advisor with End-to-end Performance Optimization

    SciTech Connect

    Zhu, Michelle M.; Wu, Chase Q.

    2013-11-07

    Next-generation eScience applications often generate large amounts of simulation, experimental, or observational data that must be shared and managed by collaborative organizations. Advanced networking technologies and services have been rapidly developed and deployed to facilitate such massive data transfer. However, these technologies and services have not been fully utilized mainly because their use typically requires significant domain knowledge and in many cases application users are even not aware of their existence. By leveraging the functionalities of an existing Network-Aware Data Movement Advisor (NADMA) utility, we propose a new Workflow-based Intelligent Network Data Movement Advisor (WINDMA) with end-to-end performance optimization for this DOE funded project. This WINDMA system integrates three major components: resource discovery, data movement, and status monitoring, and supports the sharing of common data movement workflows through account and database management. This system provides a web interface and interacts with existing data/space management and discovery services such as Storage Resource Management, transport methods such as GridFTP and GlobusOnline, and network resource provisioning brokers such as ION and OSCARS. We demonstrate the efficacy of the proposed transport-support workflow system in several use cases based on its implementation and deployment in DOE wide-area networks.

  19. End to End Digitisation and Analysis of Three-Dimensional Coral Models, from Communities to Corallites

    PubMed Central

    Gutierrez-Heredia, Luis; Benzoni, Francesca; Murphy, Emma; Reynaud, Emmanuel G.

    2016-01-01

    Coral reefs hosts nearly 25% of all marine species and provide food sources for half a billion people worldwide while only a very small percentage have been surveyed. Advances in technology and processing along with affordable underwater cameras and Internet availability gives us the possibility to provide tools and softwares to survey entire coral reefs. Holistic ecological analyses of corals require not only the community view (10s to 100s of meters), but also the single colony analysis as well as corallite identification. As corals are three-dimensional, classical approaches to determine percent cover and structural complexity across spatial scales are inefficient, time-consuming and limited to experts. Here we propose an end-to-end approach to estimate these parameters using low-cost equipment (GoPro, Canon) and freeware (123D Catch, Meshmixer and Netfabb), allowing every community to participate in surveys and monitoring of their coral ecosystem. We demonstrate our approach on 9 species of underwater colonies in ranging size and morphology. 3D models of underwater colonies, fresh samples and bleached skeletons with high quality texture mapping and detailed topographic morphology were produced, and Surface Area and Volume measurements (parameters widely used for ecological and coral health studies) were calculated and analysed. Moreover, we integrated collected sample models with micro-photogrammetry models of individual corallites to aid identification and colony and polyp scale analysis. PMID:26901845

  20. Characterisation of residual ionospheric errors in bending angles using GNSS RO end-to-end simulations

    NASA Astrophysics Data System (ADS)

    Liu, C. L.; Kirchengast, G.; Zhang, K. F.; Norman, R.; Li, Y.; Zhang, S. C.; Carter, B.; Fritzer, J.; Schwaerz, M.; Choy, S. L.; Wu, S. Q.; Tan, Z. X.

    2013-09-01

    Global Navigation Satellite System (GNSS) radio occultation (RO) is an innovative meteorological remote sensing technique for measuring atmospheric parameters such as refractivity, temperature, water vapour and pressure for the improvement of numerical weather prediction (NWP) and global climate monitoring (GCM). GNSS RO has many unique characteristics including global coverage, long-term stability of observations, as well as high accuracy and high vertical resolution of the derived atmospheric profiles. One of the main error sources in GNSS RO observations that significantly affect the accuracy of the derived atmospheric parameters in the stratosphere is the ionospheric error. In order to mitigate the effect of this error, the linear ionospheric correction approach for dual-frequency GNSS RO observations is commonly used. However, the residual ionospheric errors (RIEs) can be still significant, especially when large ionospheric disturbances occur and prevail such as during the periods of active space weather. In this study, the RIEs were investigated under different local time, propagation direction and solar activity conditions and their effects on RO bending angles are characterised using end-to-end simulations. A three-step simulation study was designed to investigate the characteristics of the RIEs through comparing the bending angles with and without the effects of the RIEs. This research forms an important step forward in improving the accuracy of the atmospheric profiles derived from the GNSS RO technique.

  1. Advanced end-to-end fiber optic sensing systems for demanding environments

    NASA Astrophysics Data System (ADS)

    Black, Richard J.; Moslehi, Behzad

    2010-09-01

    Optical fibers are small-in-diameter, light-in-weight, electromagnetic-interference immune, electrically passive, chemically inert, flexible, embeddable into different materials, and distributed-sensing enabling, and can be temperature and radiation tolerant. With appropriate processing and/or packaging, they can be very robust and well suited to demanding environments. In this paper, we review a range of complete end-to-end fiber optic sensor systems that IFOS has developed comprising not only (1) packaged sensors and mechanisms for integration with demanding environments, but (2) ruggedized sensor interrogators, and (3) intelligent decision aid algorithms software systems. We examine the following examples: " Fiber Bragg Grating (FBG) optical sensors systems supporting arrays of environmentally conditioned multiplexed FBG point sensors on single or multiple optical fibers: In conjunction with advanced signal processing, decision aid algorithms and reasoners, FBG sensor based structural health monitoring (SHM) systems are expected to play an increasing role in extending the life and reducing costs of new generations of aerospace systems. Further, FBG based structural state sensing systems have the potential to considerably enhance the performance of dynamic structures interacting with their environment (including jet aircraft, unmanned aerial vehicles (UAVs), and medical or extravehicular space robots). " Raman based distributed temperature sensing systems: The complete length of optical fiber acts as a very long distributed sensor which may be placed down an oil well or wrapped around a cryogenic tank.

  2. End to End Digitisation and Analysis of Three-Dimensional Coral Models, from Communities to Corallites.

    PubMed

    Gutierrez-Heredia, Luis; Benzoni, Francesca; Murphy, Emma; Reynaud, Emmanuel G

    2016-01-01

    Coral reefs hosts nearly 25% of all marine species and provide food sources for half a billion people worldwide while only a very small percentage have been surveyed. Advances in technology and processing along with affordable underwater cameras and Internet availability gives us the possibility to provide tools and softwares to survey entire coral reefs. Holistic ecological analyses of corals require not only the community view (10s to 100s of meters), but also the single colony analysis as well as corallite identification. As corals are three-dimensional, classical approaches to determine percent cover and structural complexity across spatial scales are inefficient, time-consuming and limited to experts. Here we propose an end-to-end approach to estimate these parameters using low-cost equipment (GoPro, Canon) and freeware (123D Catch, Meshmixer and Netfabb), allowing every community to participate in surveys and monitoring of their coral ecosystem. We demonstrate our approach on 9 species of underwater colonies in ranging size and morphology. 3D models of underwater colonies, fresh samples and bleached skeletons with high quality texture mapping and detailed topographic morphology were produced, and Surface Area and Volume measurements (parameters widely used for ecological and coral health studies) were calculated and analysed. Moreover, we integrated collected sample models with micro-photogrammetry models of individual corallites to aid identification and colony and polyp scale analysis.

  3. End-to-End Information System design at the NASA Jet Propulsion Laboratory

    NASA Technical Reports Server (NTRS)

    Hooke, A. J.

    1978-01-01

    Recognizing a pressing need of the 1980s to optimize the two-way flow of information between a ground-based user and a remote space-based sensor, an end-to-end approach to the design of information systems has been adopted at the Jet Propulsion Laboratory. The objectives of this effort are to ensure that all flight projects adequately cope with information flow problems at an early stage of system design, and that cost-effective, multi-mission capabilities are developed when capital investments are made in supporting elements. The paper reviews the End-to-End Information System (EEIS) activity at the Laboratory, and notes the ties to the NASA End-to-End Data System program.

  4. An end-to-end communications architecture for condition-based maintenance applications

    NASA Astrophysics Data System (ADS)

    Kroculick, Joseph

    2014-06-01

    This paper explores challenges in implementing an end-to-end communications architecture for Condition-Based Maintenance Plus (CBM+) data transmission which aligns with the Army's Network Modernization Strategy. The Army's Network Modernization strategy is based on rolling out network capabilities which connect the smallest unit and Soldier level to enterprise systems. CBM+ is a continuous improvement initiative over the life cycle of a weapon system or equipment to improve the reliability and maintenance effectiveness of Department of Defense (DoD) systems. CBM+ depends on the collection, processing and transport of large volumes of data. An important capability that enables CBM+ is an end-to-end network architecture that enables data to be uploaded from the platform at the tactical level to enterprise data analysis tools. To connect end-to-end maintenance processes in the Army's supply chain, a CBM+ network capability can be developed from available network capabilities.

  5. End-to-End Information System design at the NASA Jet Propulsion Laboratory

    NASA Technical Reports Server (NTRS)

    Hooke, A. J.

    1978-01-01

    Recognizing a pressing need of the 1980s to optimize the two-way flow of information between a ground-based user and a remote space-based sensor, an end-to-end approach to the design of information systems has been adopted at the Jet Propulsion Laboratory. The objectives of this effort are to ensure that all flight projects adequately cope with information flow problems at an early stage of system design, and that cost-effective, multi-mission capabilities are developed when capital investments are made in supporting elements. The paper reviews the End-to-End Information System (EEIS) activity at the Laboratory, and notes the ties to the NASA End-to-End Data System program.

  6. Measuring End-To-End Bandwidth with Iperf Using Web100

    SciTech Connect

    Cottrell, Les

    2003-04-30

    End-to-end bandwidth estimation tools like Iperf though fairly accurate are intrusive. In this paper, we describe how with an instrumented TCP stack (Web100), we can estimate the end-to-end bandwidth accurately, while consuming significantly less network bandwidth and time. We modified Iperf to use Web100 to detect the end of slow-start and estimate the end-to-end bandwidth by measuring the amount of data sent for a short period (1 second) after the slow-start, when the TCP throughput is relatively stable. We obtained bandwidth estimates differing by less than 10% when compared to running Iperf for 20 seconds, and savings in bandwidth estimation time of up to 94% and savings in network traffic of up to 92%.

  7. A Robust Method to Integrate End-to-End Mission Architecture Optimization Tools

    NASA Technical Reports Server (NTRS)

    Lugo, Rafael; Litton, Daniel; Qu, Min; Shidner, Jeremy; Powell, Richard

    2016-01-01

    End-to-end mission simulations include multiple phases of flight. For example, an end-to-end Mars mission simulation may include launch from Earth, interplanetary transit to Mars and entry, descent and landing. Each phase of flight is optimized to meet specified constraints and often depend on and impact subsequent phases. The design and optimization tools and methodologies used to combine different aspects of end-to-end framework and their impact on mission planning are presented. This work focuses on a robust implementation of a Multidisciplinary Design Analysis and Optimization (MDAO) method that offers the flexibility to quickly adapt to changing mission design requirements. Different simulations tailored to the liftoff, ascent, and atmospheric entry phases of a trajectory are integrated and optimized in the MDAO program Isight, which provides the user a graphical interface to link simulation inputs and outputs. This approach provides many advantages to mission planners, as it is easily adapted to different mission scenarios and can improve the understanding of the integrated system performance within a particular mission configuration. A Mars direct entry mission using the Space Launch System (SLS) is presented as a generic end-to-end case study. For the given launch period, the SLS launch performance is traded for improved orbit geometry alignment, resulting in an optimized a net payload that is comparable to that in the SLS Mission Planner's Guide.

  8. End-to-End Data System Architecture for the Space Station Biological Research Project

    NASA Technical Reports Server (NTRS)

    Mian, Arshad; Scimemi, Sam; Adeni, Kaiser; Picinich, Lou; Ramos, Rubin (Technical Monitor)

    1998-01-01

    The Space Station Biological Research Project (SSBRP) Is developing hardware referred to as the "facility" for providing life sciences research capability on the International Space Station. This hardware includes several biological specimen habitats, habitat holding racks, a centrifuge and a glovebox. An SSBRP end to end data system architecture has been developed to allow command and control of the facility from the ground, either with crew assistance or autonomously. The data system will be capable of handling commands, sensor data, and video from multiple cameras. The data will traverse through several onboard and ground networks and processing entities including the SSBRP and Space Station onboard and ground data systems. A large number of onboard and ground (,entities of the data system are being developed by the Space Station Program, other NASA centers and the International Partners. The SSBRP part of the system which includes the habitats, holding racks, and the ground operations center, User Operations Facility (UOF) will be developed by a multitude of geographically distributed development organizations. The SSBRP has the responsibility to define the end to end data and communications systems to make the interfaces manageable and verifiable with multiple contractors with widely varying development constraints and schedules. This paper provides an overview of the SSBRP end-to-end data system. Specifically, it describes the hardware, software and functional interactions of individual systems, and interface requirements among various entities of the end-to-end system.

  9. End-to-end ductal anastomosis in biliary reconstruction: indications and limitations.

    PubMed

    Jabłonska, Beata

    2014-08-01

    End-to-end ductal anastomosis is a physiologic biliary reconstruction that is commonly used in liver transplantation and less frequently in the surgical treatment of iatrogenic bile duct injuries. Currently, end-to-end ductal anastomosis is the biliary reconstruction of choice for liver transplantation in most adult patients. In recent years, it has also been performed for liver transplantation in children and in select patients with primary sclerosing cholangitis. The procedure is also performed in some patients with iatrogenic bile duct injuries, as it establishes physiologic bile flow. Proper digestion and absorption as well as postoperative endoscopic access are possible in patients who undergo end-to-end ductal anastomosis. It allows endoscopic diagnostic and therapeutic procedures in patients following surgery. This anastomosis is technically simple and associated with fewer early postoperative complications than the Roux-en-Y hepaticojejunostomy; however, end-to-end ductal anastomosis is not possible to perform in all patients. This review discusses the indications for and limitations of this biliary reconstruction, the technique used in liver transplantation and surgical repair of injured bile ducts, suture types and use of a T-tube.

  10. Scalable end-to-end encryption technology for supra-gigabit/second networking

    SciTech Connect

    Pierson, L.G.; Tarman, T.D.; Witzke, E.L.

    1997-05-01

    End-to-end encryption can protect proprietary information as it passes through a complex inter-city computer network, even if the intermediate systems are untrusted. This technique involves encrypting the body of computer messages while leaving network addressing and control information unencrypted for processing by intermediate network nodes. Because high speed implementations of end-to-end encryption with easy key management for standard network protocols are unavailable, this technique is not widely used today. Specifically, no end-to-end encryptors exist to protect Asynchronous Transfer Mode (ATM) traffic, nor to protect Switched Multi-megabit Data Service (SMDS), which is the first ``Broadband Integrated Services Digital Network`` (BISDN) service now being used by long distance telephone companies. This encryption technology is required for the protection of data in transit between industrial sites and central Massively Parallel Supercomputing Centers over high bandwidth, variable bit rate (BISDN) services. This research effort investigated techniques to scale end-to-end encryption technology from today`s state of the art ({approximately} 0.001 Gb/s) to 2.4 Gb/s and higher. A cryptosystem design has been developed which scales for implementation beyond SONET OC-48 (2.4Gb/s) data rates. A prototype for use with OC-3 (0.155 Gb/s) ATM variable bit rate services was developed.

  11. An end-to-end vechicle classification pipeline using vibrometry data

    NASA Astrophysics Data System (ADS)

    Smith, Ashley; Mendoza-Schrock, Olga; Kangas, Scott; Dierking, Matthew; Shaw, Arnab

    2014-06-01

    This paper evaluates and expands upon the existing end-to-end process used for vibrometry target classification and identification. A fundamental challenge in vehicle classification using vibrometry signature data is the determination of robust signal features. The methodology used in this paper involves comparing the performance of features taken from automatic speech recognition, seismology, and structural analysis work. These features provide a means to reduce the dimensionality of the data for the possibility of improved separability. The performances of different groups of features are compared to determine the best feature set for vehicle classification. Standard performance metrics are implemented to provide a method of evaluation. The contribution of this paper is to (1) thoroughly explain the time domain and frequency domain features that have been recently applied to the vehicle classification using laser-vibrometry data domain, (2) build an end-to-end classification pipeline for Aided Target Recognition (ATR) with common and easily accessible tools, and (3) apply feature selection methods to the end-to-end pipeline. The end-to-end process used here provides a structured path for accomplishing vibrometry-based target identification. This paper will compare with two studies in the public domain. The techniques utilized in this paper were utilized to analyze a small in-house database of several different vehicles.

  12. Context-driven, prescription-based personal activity classification: methodology, architecture, and end-to-end implementation.

    PubMed

    Xu, James Y; Chang, Hua-I; Chien, Chieh; Kaiser, William J; Pottie, Gregory J

    2014-05-01

    Enabling large-scale monitoring and classification of a range of motion activities is of primary importance due to the need by healthcare and fitness professionals to monitor exercises for quality and compliance. Past work has not fully addressed the unique challenges that arise from scaling. This paper presents a novel end-to-end system solution to some of these challenges. The system is built on the prescription-based context-driven activity classification methodology. First, we show that by refining the definition of context, and introducing the concept of scenarios, a prescription model can provide personalized activity monitoring. Second, through a flexible architecture constructed from interface models, we demonstrate the concept of a context-driven classifier. Context classification is achieved through a classification committee approach, and activity classification follows by means of context specific activity models. Then, the architecture is implemented in an end-to-end system featuring an Android application running on a mobile device, and a number of classifiers as core classification components. Finally, we use a series of experimental field evaluations to confirm the expected benefits of the proposed system in terms of classification accuracy, rate, and sensor operating life.

  13. Space Debris End to End Service - presentation of the DLR work plan

    NASA Astrophysics Data System (ADS)

    Alwes, D.

    2001-10-01

    Resulting from the present effort on international regulations and the compilation of guidelines to handle the space debris problem in the future, the need consists to advise spacecraft manufacturer, suppliers and operators ideally in the form of a commercially useful End to End Service. The competence of German institutions incl. research institutes and companies in the field of the space debris research and risk analyses is widely recognized and very well suitable to deal with all the corresponding aspects in Germany but also Europe-wide. It is the intention of DLR to concentrate and link all expertises and competences, available in Germany, in one coherent project. The space industry should contribute to risk analyses, reviews of guidelines and to proposals for space debris minimising design and operations of spacecraft. Preliminary proposals by several institutes and companies have been prepared and have been compiled to a consolidated work plan for a project named 'Space Debris End to End Service'. 'End to End' means the consideration of space debris mitigation aspects from begin of the design over the operational phase of a spacecraft up to the end of the mission and the following phase of the initiation of space debris avoidance measures (e.g. passivation, de- or re-orbiting). The effectiveness of such an End to End Service shall be demonstrated on a satellite project. The work packages of the project shall cover in principle the following activities: Conception of an End to End Service - Identification of national needs, the state of knowledge and information - Spacecraft requirements and mitigation measures - Re-entry analyses - Application on a pilot project, incl. system review, meteoroid and space debris modelling, hazard analyses, recommendations of measures and a cost to benefit analysis. This paper reports about the work plan and the status of the project.

  14. SU-E-T-150: End to End Tests On the First Clinical EDGETM

    SciTech Connect

    Scheib, S; Schmelzer, P; Vieira, S; Greco, C

    2014-06-01

    Purpose: To quantify the sub millimeter overall accuracy of EDGETM, the dedicated linac based SRS/SABR treatment platform from Varian, using a novel End-to-End (E2E) test phantom. Methods: The new E2E test phantom developed by Varian consists of a cube with an outer dimension of 15x15x15 cm3. The phantom is equipped with an exchangable inner cube (7×7×7 cm3) to hold radiochromic films or a tungsten ball (diameter = 5 mm) for Winston-Lutz tests. 16 ceramic balls (diameter = 5 mm) are embedded in the outer cube. Three embedded Calypso transponders allow for Calypso based monitoring. The outer surface of the phantom is tracked using the Optical Surface Monitoring System (OSMS). The phantom is positioned using kV, MV and CBCT images. A simCT of the phantom was acquired and SRS/SABR plans were treated using the new phantom on the first clinical installed EDGETM. As a first step a series of EPID based Winston-Lutz tests have been performed. As a second step the calculated dose distribution applied to the phantom was verified with radiochromic films in orthogonal planes. The measured dose distribution is compared with the calculated (Eclipse) one based on the known isocenter on both dose distributions. The geometrical shift needed to match both dose distributions is the overall accuracy and is determined using dose profiles, isodose lines or gamma pass rates (3%, 1 mm). Results: Winston-Lutz tests using the central tungsten BB demonstrated a targeting accuracy of 0.44±0.18mm for jaw (2cm × 2cm) defined 0.39±0.19mm for MLC (2cm × 2cm) defined and 0.37±0.15mm for cone (12.5 mm) defined fields. A treated patient plan (spinal metastases lesion with integrated boost) showed a dosimetric dose localization accuracy of 0.6mm. Conclusion: Geometric and dosimetric E2E tests on EDGETM, show sub-millimeter E2E targeting and dose localisation accuracy.

  15. End-to-end network models encompassing terrestrial, wireless, and satellite components

    NASA Astrophysics Data System (ADS)

    Boyarko, Chandler L.; Britton, John S.; Flores, Phil E.; Lambert, Charles B.; Pendzick, John M.; Ryan, Christopher M.; Shankman, Gordon L.; Williams, Ramon P.

    2004-08-01

    Development of network models that reflect true end-to-end architectures such as the Transformational Communications Architecture need to encompass terrestrial, wireless and satellite component to truly represent all of the complexities in a world wide communications network. Use of best-in-class tools including OPNET, Satellite Tool Kit (STK), Popkin System Architect and their well known XML-friendly definitions, such as OPNET Modeler's Data Type Description (DTD), or socket-based data transfer modules, such as STK/Connect, enable the sharing of data between applications for more rapid development of end-to-end system architectures and a more complete system design. By sharing the results of and integrating best-in-class tools we are able to (1) promote sharing of data, (2) enhance the fidelity of our results and (3) allow network and application performance to be viewed in the context of the entire enterprise and its processes.

  16. End-to-end RMS error testing on a constant bandwidth FM/FM system

    NASA Technical Reports Server (NTRS)

    Wallace, G. R.; Salter, W. E.

    1972-01-01

    End-to-end root-mean-square (rms) tests performed on a constant bandwidth FM/FM system with various settings of system parameters are reported. The testing technique employed is that of sampling, digitizing, delaying, and comparing the analog input against the sampled and digitized corresponding output. Total system error is determined by fully loading all channels with band-limited noise and conducting end-to-end rms error tests on one channel. Tests are also conducted with and without a transmission link and plots of rms errors versus receiver signal-to-noise (S/N) values are obtained. The combined effects of intermodulation, adjacent channel crosstalk, and residual system noise are determined as well as the single channel distortion of the system.

  17. Providing end-to-end QoS for multimedia applications in 3G wireless networks

    NASA Astrophysics Data System (ADS)

    Guo, Katherine; Rangarajan, Samapth; Siddiqui, M. A.; Paul, Sanjoy

    2003-11-01

    As the usage of wireless packet data services increases, wireless carriers today are faced with the challenge of offering multimedia applications with QoS requirements within current 3G data networks. End-to-end QoS requires support at the application, network, link and medium access control (MAC) layers. We discuss existing CDMA2000 network architecture and show its shortcomings that prevent supporting multiple classes of traffic at the Radio Access Network (RAN). We then propose changes in RAN within the standards framework that enable support for multiple traffic classes. In addition, we discuss how Session Initiation Protocol (SIP) can be augmented with QoS signaling for supporting end-to-end QoS. We also review state of the art scheduling algorithms at the base station and provide possible extensions to these algorithms to support different classes of traffic as well as different classes of users.

  18. Laser welding with an albumin stent: experimental ureteral end-to-end anastomosis

    NASA Astrophysics Data System (ADS)

    Xie, Hua; Shaffer, Brian S.; Prahl, Scott A.; Gregory, Kenton W.

    2000-05-01

    Porcine ureters were anastomosed using an albumin stent and diode laser in vitro. The albumin stent provided precise apposition for an end to end anastomosis and enhanced welding strength. The anastomosis seam was lasered with an 810 nm diode laser using continuous wave and pulse light through a hand-held 600 micrometer noncontact optical fiber. Tensile strength, burst pressures, operative times, total energy and thermal damaged were measured in this study. The results demonstrated that using an albumin stent to laser weld ureteral anastomoses produces strong weld strengths. The liquid albumin solder also provided satisfactory welding strength. There were no significant differences of tissue thermal damage between the albumin stent alone, liquid solder alone and both combination groups. Thermal damage to tissue depended on laser setting and energy. This study determined the appropriate laser setting parameters to perform in vivo ureteral end to end anastomosis.

  19. Sutureless end-to-end ureteral anastomosis using a new albumin stent and diode laser

    NASA Astrophysics Data System (ADS)

    Xie, Hua; Shaffer, Brian S.; Prahl, Scott A.; Gregory, Kenton W.

    1999-09-01

    Sutureless end to end ureteral anastomoses was successfully constructed in acute and chronic experiments. A photothermal sensitive hydrolyzable (PSH) albumin stent played roles as solder and intraluminal supporter to adhesion and position the anastomosed ureter by end to end fashion. The anastomosis seam was lased with 810 nm diode laser energy supplied through hand- held 600 micrometers noncontact optical fiber. A continuous 1 watt wave of power was applied for laser anastomosis. Integrity, welding strength, bursting pressures of anastomosis and histological reaction, and radiological phenomena were compared to those of anastomoses constructed using a liquidity soldering technique. The acute results of two methods were equivalent at welding strengths, but the liquid soldering showed more energy consumption. At chronic study, the radiological and histological studies were performed to evaluate the complications of the anastomosis. Excellent heating and varied degrees of complications were observed. We conclude that PSH stent showed great promise for ureteral anastomosis using laser welding.

  20. Radiometry simulation within the end-to-end simulation tool SENSOR

    NASA Astrophysics Data System (ADS)

    Wiest, Lorenz; Boerner, Anko

    2001-02-01

    12 An end-to-end simulation is a valuable tool for sensor system design, development, optimization, testing, and calibration. This contribution describes the radiometry module of the end-to-end simulation tool SENSOR. It features MODTRAN 4.0-based look up tables in conjunction with a cache-based multilinear interpolation algorithm to speed up radiometry calculations. It employs a linear reflectance parameterization to reduce look up table size, considers effects due to the topology of a digital elevation model (surface slope, sky view factor) and uses a reflectance class feature map to assign Lambertian and BRDF reflectance properties to the digital elevation model. The overall consistency of the radiometry part is demonstrated by good agreement between ATCOR 4-retrieved reflectance spectra of a simulated digital image cube and the original reflectance spectra used to simulate this image data cube.

  1. End-to-end calculation of the radiation characteristics of VVER-1000 spent fuel assemblies

    SciTech Connect

    Linge, I. I.; Mitenkova, E. F. Novikov, N. V.

    2012-12-15

    The results of end-to-end calculation of the radiation characteristics of VVER-1000 spent nuclear fuel are presented. Details of formation of neutron and gamma-radiation sources are analyzed. Distributed sources of different types of radiation are considered. A comparative analysis of calculated radiation characteristics is performed with the use of nuclear data from different ENDF/B and EAF files and ANSI/ANS and ICRP standards.

  2. The International Space Station Alpha (ISSA) End-to-End On-Orbit Maintenance Process Flow

    NASA Technical Reports Server (NTRS)

    Zingrebe, Kenneth W., II

    1995-01-01

    As a tool for construction and refinement of the on-orbit maintenance system to sustain the International Space Station Alpha (ISSA), the Mission Operations Directorate (MOD) developed an end to-end on-orbit maintenance process flow. This paper discusses and demonstrates that process flow. This tool is being used by MOD to identify areas which require further work in preparation for MOD's role in the conduct of on-orbit maintenance operations.

  3. Severing and end-to-end annealing of neurofilaments in neurons

    PubMed Central

    Uchida, Atsuko; Çolakoğlu, Gülsen; Wang, Lina; Monsma, Paula C.; Brown, Anthony

    2013-01-01

    We have shown previously that neurofilaments and vimentin filaments expressed in nonneuronal cell lines can lengthen by joining ends in a process known as “end-to-end annealing.” To test if this also occurs for neurofilaments in neurons, we transfected cultured rat cortical neurons with fluorescent neurofilament fusion proteins and then used photoconversion or photoactivation strategies to create distinct populations of red and green fluorescent filaments. Within several hours we observed the appearance of chimeric filaments consisting of alternating red and green segments, which is indicative of end-to-end annealing of red and green filaments. However, the appearance of these chimeric filaments was accompanied by a gradual fragmentation of the red and green filament segments, which is indicative of severing. Over time we observed a progressive increase in the number of red–green junctions along the filaments accompanied by a progressive decrease in the average length of the alternating red and green fluorescent segments that comprised those filaments, suggesting a dynamic cycle of severing and end-to-end-annealing. Time-lapse imaging of the axonal transport of chimeric filaments demonstrated that the red and green segments moved together, confirming that they were indeed part of the same filament. Moreover, in several instances, we also were able to capture annealing and severing events live in time-lapse movies. We propose that the length of intermediate filaments in cells is regulated by the opposing actions of severing and end-to-end annealing, and we speculate that this regulatory mechanism may influence neurofilament transport within axons. PMID:23821747

  4. CHEETAH: circuit-switched high-speed end-to-end transport architecture

    NASA Astrophysics Data System (ADS)

    Veeraraghavan, Malathi; Zheng, Xuan; Lee, Hyuk; Gardner, M.; Feng, Wuchun

    2003-10-01

    Leveraging the dominance of Ethernet in LANs and SONET/SDH in MANs and WANs, we propose a service called CHEETAH (Circuit-switched High-speed End-to-End Transport ArcHitecture). The service concept is to provide end hosts with high-speed, end-to-end circuit connectivity on a call-by-call shared basis, where a "circuit" consists of Ethernet segments at the ends that are mapped into Ethernet-over-SONET long-distance circuits. This paper focuses on the file-transfer application for such circuits. For this application, the CHEETAH service is proposed as an add-on to the primary Internet access service already in place for enterprise hosts. This allows an end host that is sending a file to first attempt setting up an end-to-end Ethernet/EoS circuit, and if rejected, fall back to the TCP/IP path. If the circuit setup is successful, the end host will enjoy a much shorter file-transfer delay than on the TCP/IP path. To determine the conditions under which an end host with access to the CHEETAH service should attempt circuit setup, we analyze mean file-transfer delays as a function of call blocking probability in the circuit-switched network, probability of packet loss in the IP network, round-trip times, link rates, and so on.

  5. Sulfate-Mediated End-to-End Assembly of Gold Nanorods.

    PubMed

    Abtahi, S M H; Burrows, Nathan D; Idesis, Fred A; Murphy, Catherine J; Saleh, Navid B; Vikesland, Peter J

    2017-02-14

    There is interest in the controlled aggregation of gold nanorods (GNRs) for the production of extended nanoassemblies. Prior studies have relied upon chemical modification of the GNR surface to achieve a desired final aggregate structure. Herein we illustrate that control of electrolyte composition can facilitate end-to-end assembly of cetyltrimethylammonium-bromide-coated (CTAB) GNRs. By adjusting either the sulfate anion concentration or the exposure time it is possible to connect GNRs in chain-like assemblies. In contrast, end-to-end assembly was not observed in control experiments using monovalent chloride salts. We attribute the end-to-end assembly to the localized association of sulfate with exposed quaternary ammonium head groups of CTAB at the nanorod tip. To quantify the assembly kinetics, visible-near-infrared extinction spectra were collected over a predetermined time period, and the colloidal behavior of the GNR suspensions was interpreted using plasmon band analysis. Transmission electron microscopy and atomic force microscopy results support the conclusions reached via plasmon band analysis, and the colloidal behavior is consistent with Derjaguin-Landau-Verwey-Overbeek theory.

  6. An end-to-end approach to developing biological and chemical detector requirements

    NASA Astrophysics Data System (ADS)

    Teclemariam, Nerayo P.; Purvis, Liston K.; Foltz, Greg W.; West, Todd; Edwards, Donna M.; Fruetel, Julia A.; Gleason, Nathaniel J.

    2009-05-01

    Effective defense against chemical and biological threats requires an "end-to-end" strategy that encompasses the entire problem space, from threat assessment and target hardening to response planning and recovery. A key element of the strategy is the definition of appropriate system requirements for surveillance and detection of threat agents. Our end-to-end approach to venue chem/bio defense is captured in the Facilities Weapons of Mass Destruction Decision Analysis Capability (FacDAC), an integrated system-of-systems toolset that can be used to generate requirements across all stages of detector development. For example, in the early stage of detector development the approach can be used to develop performance targets (e.g., sensitivity, selectivity, false positive rate) to provide guidance on what technologies to pursue. In the development phase, after a detector technology has been selected, the approach can aid in determining performance trade-offs and down-selection of competing technologies. During the application stage, the approach can be employed to design optimal defensive architectures that make the best use of available technology to maximize system performance. This presentation will discuss the end-to-end approach to defining detector requirements and demonstrate the capabilities of the FacDAC toolset using examples from a number of studies for the Department of Homeland Security.

  7. Direct muscle neurotization after end-to end and end-to-side neurorrhaphy

    PubMed Central

    Papalia, Igor; Ronchi, Giulia; Muratori, Luisa; Mazzucco, Alessandra; Magaudda, Ludovico; Geuna, Stefano

    2012-01-01

    The need for the continuous research of new tools for improving motor function recovery after nerve injury is justified by the still often unsatisfactory clinical outcome in these patients. It has been previously shown that the combined use of two reconstructive techniques, namely end-to-side neurorrhaphy and direct muscle neurotization in the rat hindlimb model, can lead to good results in terms of skeletal muscle reinnervation. Here we show that, in the rat forelimb model, the combined use of direct muscle neurotization with either end-to-end or end-to-side neurorrhaphy to reinnervate the denervated flexor digitorum muscles, leads to muscle atrophy prevention over a long postoperative time lapse (10 months). By contrast, very little motor recovery (in case of end-to-end neurorrhaphy) and almost no motor recovery (in case of end-to-side neurorrhaphy) were observed in the grasping activity controlled by flexor digitorum muscles. It can thus be concluded that, at least in the rat, direct muscle neurotization after both end-to-end and end-to-side neurorrhaphy represents a good strategy for preventing denervation-related muscle atrophy but not for regaining the lost motor function. PMID:25538749

  8. End-to-end System Performance Simulation: A Data-Centric Approach

    NASA Astrophysics Data System (ADS)

    Guillaume, Arnaud; Laffitte de Petit, Jean-Luc; Auberger, Xavier

    2013-08-01

    In the early times of space industry, the feasibility of Earth observation missions was directly driven by what could be achieved by the satellite. It was clear to everyone that the ground segment would be able to deal with the small amount of data sent by the payload. Over the years, the amounts of data processed by the spacecrafts have been increasing drastically, leading to put more and more constraints on the ground segment performances - and in particular on timeliness. Nowadays, many space systems require high data throughputs and short response times, with information coming from multiple sources and involving complex algorithms. It has become necessary to perform thorough end-to-end analyses of the full system in order to optimise its cost and efficiency, but even sometimes to assess the feasibility of the mission. This paper presents a novel framework developed by Astrium Satellites in order to meet these needs of timeliness evaluation and optimisation. This framework, named ETOS (for “End-to-end Timeliness Optimisation of Space systems”), provides a modelling process with associated tools, models and GUIs. These are integrated thanks to a common data model and suitable adapters, with the aim of building suitable space systems simulators of the full end-to-end chain. A big challenge of such environment is to integrate heterogeneous tools (each one being well-adapted to part of the chain) into a relevant timeliness simulation.

  9. On the importance of risk knowledge for an end-to-end tsunami early warning system

    NASA Astrophysics Data System (ADS)

    Post, Joachim; Strunz, Günter; Riedlinger, Torsten; Mück, Matthias; Wegscheider, Stephanie; Zosseder, Kai; Steinmetz, Tilmann; Gebert, Niklas; Anwar, Herryal

    2010-05-01

    context has been worked out. The generated results contribute significantly in the fields of (1) warning decision and warning levels, (2) warning dissemination and warning message content, (3) early warning chain planning, (4) increasing response capabilities and protective systems, (5) emergency relief and (6) enhancing communities' awareness and preparedness towards tsunami threats. Additionally examples will be given on the potentials of an operational use of risk information in early warning systems as first experiences exist for the tsunami early warning center in Jakarta, Indonesia. Beside this the importance of linking national level early warning information with tsunami risk information available at the local level (e.g. linking warning message information on expected intensity with respective tsunami hazard zone maps at community level for effective evacuation) will be demonstrated through experiences gained in three pilot areas in Indonesia. The presentation seeks to provide new insights on benefits using risk information in early warning and will provide further evidence that practical use of risk information is an important and indispensable component of end-to-end early warning.

  10. SensorKit: An End-to-End Solution for Environmental Sensor Networking

    NASA Astrophysics Data System (ADS)

    Silva, F.; Graham, E.; Deschon, A.; Lam, Y.; Goldman, J.; Wroclawski, J.; Kaiser, W.; Benzel, T.

    2008-12-01

    Modern day sensor network technology has shown great promise to transform environmental data collection. However, despite the promise, these systems have remained the purview of the engineers and computer scientists who design them rather than a useful tool for the environmental scientists who need them. SensorKit is conceived of as a way to make wireless sensor networks accessible to The People: it is an advanced, powerful tool for sensor data collection that does not require advanced technological know-how. We are aiming to make wireless sensor networks for environmental science as simple as setting up a standard home computer network by providing simple, tested configurations of commercially-available hardware, free and easy-to-use software, and step-by-step tutorials. We designed and built SensorKit using a simplicity-through-sophistication approach, supplying users a powerful sensor to database end-to-end system with a simple and intuitive user interface. Our objective in building SensorKit was to make the prospect of using environmental sensor networks as simple as possible. We built SensorKit from off the shelf hardware components, using the Compact RIO platform from National Instruments for data acquisition due to its modular architecture and flexibility to support a large number of sensor types. In SensorKit, we support various types of analog, digital and networked sensors. Our modular software architecture allows us to abstract sensor details and provide users a common way to acquire data and to command different types of sensors. SensorKit is built on top of the Sensor Processing and Acquisition Network (SPAN), a modular framework for acquiring data in the field, moving it reliably to the scientist institution, and storing it in an easily-accessible database. SPAN allows real-time access to the data in the field by providing various options for long haul communication, such as cellular and satellite links. Our system also features reliable data storage

  11. Optimizing End-to-End Big Data Transfers over Terabits Network Infrastructure

    SciTech Connect

    Kim, Youngjae; Atchley, Scott; Vallee, Geoffroy R.; Lee, Sangkeun; Shipman, Galen M.

    2016-04-05

    While future terabit networks hold the promise of significantly improving big-data motion among geographically distributed data centers, significant challenges must be overcome even on today's 100 gigabit networks to realize end-to-end performance. Multiple bottlenecks exist along the end-to-end path from source to sink, for instance, the data storage infrastructure at both the source and sink and its interplay with the wide-area network are increasingly the bottleneck to achieving high performance. In this study, we identify the issues that lead to congestion on the path of an end-to-end data transfer in the terabit network environment, and we present a new bulk data movement framework for terabit networks, called LADS. LADS exploits the underlying storage layout at each endpoint to maximize throughput without negatively impacting the performance of shared storage resources for other users. LADS also uses the Common Communication Interface (CCI) in lieu of the sockets interface to benefit from hardware-level zero-copy, and operating system bypass capabilities when available. It can further improve data transfer performance under congestion on the end systems using buffering at the source using flash storage. With our evaluations, we show that LADS can avoid congested storage elements within the shared storage resource, improving input/output bandwidth, and data transfer rates across the high speed networks. We also investigate the performance degradation problems of LADS due to I/O contention on the parallel file system (PFS), when multiple LADS tools share the PFS. We design and evaluate a meta-scheduler to coordinate multiple I/O streams while sharing the PFS, to minimize the I/O contention on the PFS. Finally, with our evaluations, we observe that LADS with meta-scheduling can further improve the performance by up to 14 percent relative to LADS without meta-scheduling.

  12. End-to-end modeling: a new modular and flexible approach

    NASA Astrophysics Data System (ADS)

    Genoni, M.; Riva, M.; Landoni, M.; Pariani, G.

    2016-08-01

    In this paper we present an innovative philosophy to develop the End-to-End model for astronomical observation projects, i.e. the architecture which allows physical modeling of the whole system from the light source to the reduced data. This alternative philosophy foresees the development of the physical model of the different modules, which compose the entire End-to-End system, directly during the project design phase. This approach is strongly characterized by modularity and flexibility; these aspects will be of relevant importance in the next generation astronomical observation projects like E-ELT (European Extremely Large Telescope) because of the high complexity and long-time design and development. With this approach it will be possible to keep the whole system and its different modules efficiently under control during every project phase and to exploit a reliable tool at a system engineering level to evaluate the effects on the final performance both of the main parameters and of different instrument architectures and technologies. This philosophy will be important to allow scientific community to perform in advance simulations and tests on the scientific drivers. This will translate in a continuous feedback to the (system) design process with a resulting improvement in the effectively achievable scientific goals and consistent tool for efficiently planning observation proposals and programs. We present the application case for this End-to-End modeling technique, which is the high resolution spectrograph at the E-ELT (E-ELT HIRES). In particular we present the definition of the system modular architecture, describing the interface parameters of the modules.

  13. Optimizing End-to-End Big Data Transfers over Terabits Network Infrastructure

    DOE PAGES

    Kim, Youngjae; Atchley, Scott; Vallee, Geoffroy R.; ...

    2016-04-05

    While future terabit networks hold the promise of significantly improving big-data motion among geographically distributed data centers, significant challenges must be overcome even on today's 100 gigabit networks to realize end-to-end performance. Multiple bottlenecks exist along the end-to-end path from source to sink, for instance, the data storage infrastructure at both the source and sink and its interplay with the wide-area network are increasingly the bottleneck to achieving high performance. In this study, we identify the issues that lead to congestion on the path of an end-to-end data transfer in the terabit network environment, and we present a new bulkmore » data movement framework for terabit networks, called LADS. LADS exploits the underlying storage layout at each endpoint to maximize throughput without negatively impacting the performance of shared storage resources for other users. LADS also uses the Common Communication Interface (CCI) in lieu of the sockets interface to benefit from hardware-level zero-copy, and operating system bypass capabilities when available. It can further improve data transfer performance under congestion on the end systems using buffering at the source using flash storage. With our evaluations, we show that LADS can avoid congested storage elements within the shared storage resource, improving input/output bandwidth, and data transfer rates across the high speed networks. We also investigate the performance degradation problems of LADS due to I/O contention on the parallel file system (PFS), when multiple LADS tools share the PFS. We design and evaluate a meta-scheduler to coordinate multiple I/O streams while sharing the PFS, to minimize the I/O contention on the PFS. Finally, with our evaluations, we observe that LADS with meta-scheduling can further improve the performance by up to 14 percent relative to LADS without meta-scheduling.« less

  14. End-to-End Comparative Attention Networks for Person Re-Identification

    NASA Astrophysics Data System (ADS)

    Liu, Hao; Feng, Jiashi; Qi, Meibin; Jiang, Jianguo; Yan, Shuicheng

    2017-07-01

    Person re-identification across disjoint camera views has been widely applied in video surveillance yet it is still a challenging problem. One of the major challenges lies in the lack of spatial and temporal cues, which makes it difficult to deal with large variations of lighting conditions, viewing angles, body poses and occlusions. Recently, several deep learning based person re-identification approaches have been proposed and achieved remarkable performance. However, most of those approaches extract discriminative features from the whole frame at one glimpse without differentiating various parts of the persons to identify. It is essentially important to examine multiple highly discriminative local regions of the person images in details through multiple glimpses for dealing with the large appearance variance. In this paper, we propose a new soft attention based model, i.e., the end to-end Comparative Attention Network (CAN), specifically tailored for the task of person re-identification. The end-to-end CAN learns to selectively focus on parts of pairs of person images after taking a few glimpses of them and adaptively comparing their appearance. The CAN model is able to learn which parts of images are relevant for discerning persons and automatically integrates information from different parts to determine whether a pair of images belongs to the same person. In other words, our proposed CAN model simulates the human perception process to verify whether two images are from the same person. Extensive experiments on three benchmark person re-identification datasets, including CUHK01, CHUHK03 and Market-1501, clearly demonstrate that our proposed end-to-end CAN for person re-identification outperforms well established baselines significantly and offer new state-of-the-art performance.

  15. LWIR scene simulator developed for end-to-end performance evaluation of focal planes

    NASA Technical Reports Server (NTRS)

    Thompson, Niels A.; Bowser, William M.; Song, Sung H.; Skiff, Laura T.; Powell, William W.; Romero, Charles

    1992-01-01

    The development of a long-wave infrared optical simulator facilitates evaluation of the end-to-end performance of long wavelength infrared (LWIR) focal plane arrays (FPAs) in a system-like environment. This simulator provides selectable structured scene inputs to a focal plane module or array. Background irradiances as low as 10 exp 10 photons/sq cm s are achievable when the simulator is cooled with liquid helium. The optical simulator can generate single or multiple targets of controllable intensities, and uniform or structured background irradiances. The infrared scenes can be viewed in a stationary mode or dynamically scanned across the focal plane.

  16. Screening California Current fishery management scenarios using the Atlantis end-to-end ecosystem model

    NASA Astrophysics Data System (ADS)

    Kaplan, Isaac C.; Horne, Peter J.; Levin, Phillip S.

    2012-09-01

    End-to-end marine ecosystem models link climate and oceanography to the food web and human activities. These models can be used as forecasting tools, to strategically evaluate management options and to support ecosystem-based management. Here we report the results of such forecasts in the California Current, using an Atlantis end-to-end model. We worked collaboratively with fishery managers at NOAA’s regional offices and staff at the National Marine Sanctuaries (NMS) to explore the impact of fishery policies on management objectives at different spatial scales, from single Marine Sanctuaries to the entire Northern California Current. In addition to examining Status Quo management, we explored the consequences of several gear switching and spatial management scenarios. Of the scenarios that involved large scale management changes, no single scenario maximized all performance metrics. Any policy choice would involve trade-offs between stakeholder groups and policy goals. For example, a coast-wide 25% gear shift from trawl to pot or longline appeared to be one possible compromise between an increase in spatial management (which sacrificed revenue) and scenarios such as the one consolidating bottom impacts to deeper areas (which did not perform substantially differently from Status Quo). Judged on a coast-wide scale, most of the scenarios that involved minor or local management changes (e.g. within Monterey Bay NMS only) yielded results similar to Status Quo. When impacts did occur in these cases, they often involved local interactions that were difficult to predict a priori based solely on fishing patterns. However, judged on the local scale, deviation from Status Quo did emerge, particularly for metrics related to stationary species or variables (i.e. habitat and local metrics of landed value or bycatch). We also found that isolated management actions within Monterey Bay NMS would cause local fishers to pay a cost for conservation, in terms of reductions in landed

  17. End-to-end science from major facilities: does the VO have a role?

    NASA Astrophysics Data System (ADS)

    Gilmore, Gerard F.

    2007-08-01

    The Virtual Observatory provides a natural solution to the existence problem in communications: how can one ask a question of another unless you know the other exists? Many think e-mail from apparent strangers, e-blogs, etc., suggest there is no shortage of possible such solutions. In that context, is the Virtual Observatory in fact the necessary and desirable part of the solution? Specifically, does the VO necessarily play a critical role in delivering end-to-end facility science, from ideas, through proposals, resources/facilities, to distributed, reviewed, knowledge? If not, what else needs to be added?

  18. The first sutureless, laser-welded, end-to-end bowel anastomosis.

    PubMed

    Sauer, J S; Hinshaw, J R; McGuire, K P

    1989-01-01

    The use of laser energy to weld together tissue offers great promise in the expanding field of laser surgery. The published results of laser welding intestinal tissue have, to date, been limited to the successful laser closures of small enterotomies. This is the first report of using laser energy alone to create an end-to-end small bowel anastomosis. A biocompatible, water-soluble, intraluminal stent was employed during the laser welding of this sutureless, stapleless ileal anastomosis in a rabbit model. Excellent recovery and healing were observed. The rapidity, ease, and potential for full precision automation of laser welding mandates further research.

  19. Deriving comprehensive error breakdown for wide field adaptive optics systems using end-to-end simulations

    NASA Astrophysics Data System (ADS)

    Ferreira, F.; Gendron, E.; Rousset, G.; Gratadour, D.

    2016-07-01

    The future European Extremely Large Telescope (E-ELT) adaptive optics (AO) systems will aim at wide field correction and large sky coverage. Their performance will be improved by using post processing techniques, such as point spread function (PSF) deconvolution. The PSF estimation involves characterization of the different error sources in the AO system. Such error contributors are difficult to estimate: simulation tools are a good way to do that. We have developed in COMPASS (COMputing Platform for Adaptive opticS Systems), an end-to-end simulation tool using GPU (Graphics Processing Unit) acceleration, an estimation tool that provides a comprehensive error budget by the outputs of a single simulation run.

  20. Satellite/Terrestrial Networks: End-to-End Communication Interoperability Quality of Service Experiments

    NASA Technical Reports Server (NTRS)

    Ivancic, William D.

    1998-01-01

    Various issues associated with satellite/terrestrial end-to-end communication interoperability are presented in viewgraph form. Specific topics include: 1) Quality of service; 2) ATM performance characteristics; 3) MPEG-2 transport stream mapping to AAL-5; 4) Observation and discussion of compressed video tests over ATM; 5) Digital video over satellites status; 6) Satellite link configurations; 7) MPEG-2 over ATM with binomial errors; 8) MPEG-2 over ATM channel characteristics; 8) MPEG-2 over ATM over emulated satellites; 9) MPEG-2 transport stream with errors; and a 10) Dual decoder test.

  1. End-to-end planning and scheduling systems technology for space operations

    NASA Astrophysics Data System (ADS)

    Moe, Karen L.

    1992-08-01

    Consideration is given to planning and scheduling operations concepts from an end-to-end perspective, through both mission operations and institutional support functions. An operations concept is proposed which is based on a flexible request language used to state resource requirements and mission constraints to a scheduling system. The language has the potential to evolve into an international standard for exchanging service request information on international space networks. The key benefit of the flexible scheduling request concept is the shift of a significant conflict resolution effort from humans to computers, reducing the time for generating a week's worth of schedules to hours instead of days.

  2. LWIR scene simulator developed for end-to-end performance evaluation of focal planes

    NASA Technical Reports Server (NTRS)

    Thompson, Niels A.; Bowser, William M.; Song, Sung H.; Skiff, Laura T.; Powell, William W.; Romero, Charles

    1992-01-01

    The development of a long-wave infrared optical simulator facilitates evaluation of the end-to-end performance of long wavelength infrared (LWIR) focal plane arrays (FPAs) in a system-like environment. This simulator provides selectable structured scene inputs to a focal plane module or array. Background irradiances as low as 10 exp 10 photons/sq cm s are achievable when the simulator is cooled with liquid helium. The optical simulator can generate single or multiple targets of controllable intensities, and uniform or structured background irradiances. The infrared scenes can be viewed in a stationary mode or dynamically scanned across the focal plane.

  3. End-to-End Assessment of a Large Aperture Segmented Ultraviolet Optical Infrared (UVOIR) Telescope Architecture

    NASA Technical Reports Server (NTRS)

    Feinberg, Lee; Bolcar, Matt; Liu, Alice; Guyon, Olivier; Stark,Chris; Arenberg, Jon

    2016-01-01

    Key challenges of a future large aperture, segmented Ultraviolet Optical Infrared (UVOIR) Telescope capable of performing a spectroscopic survey of hundreds of Exoplanets will be sufficient stability to achieve 10-10 contrast measurements and sufficient throughput and sensitivity for high yield Exo-Earth spectroscopic detection. Our team has collectively assessed an optimized end to end architecture including a high throughput coronagraph capable of working with a segmented telescope, a cost-effective and heritage based stable segmented telescope, a control architecture that minimizes the amount of new technologies, and an Exo-Earth yield assessment to evaluate potential performance.

  4. End-to-end planning and scheduling systems technology for space operations

    NASA Technical Reports Server (NTRS)

    Moe, Karen L.

    1992-01-01

    Consideration is given to planning and scheduling operations concepts from an end-to-end perspective, through both mission operations and institutional support functions. An operations concept is proposed which is based on a flexible request language used to state resource requirements and mission constraints to a scheduling system. The language has the potential to evolve into an international standard for exchanging service request information on international space networks. The key benefit of the flexible scheduling request concept is the shift of a significant conflict resolution effort from humans to computers, reducing the time for generating a week's worth of schedules to hours instead of days.

  5. Solving Large Sparse Linear Systems in End-to-end Accelerator Structure Simulations

    SciTech Connect

    Lee, L

    2004-01-23

    This paper presents a case study of solving very large sparse linear systems in end-to-end accelerator structure simulations. Both direct solvers and iterative solvers are investigated. A parallel multilevel preconditioner based on hierarchical finite element basis functions is considered and has been implemented to accelerate the convergence of iterative solvers. A linear system with matrix size 93,147,736 and with 3,964,961,944 non-zeros from 3D electromagnetic finite element discretization has been solved in less than 8 minutes with 1024 CPUs on the NERSC IBM SP. The resource utilization as well as the application performance for these solvers is discussed.

  6. Ocean Acidification Scientific Data Stewardship: An approach for end-to-end data management and integration

    NASA Astrophysics Data System (ADS)

    Arzayus, K. M.; Garcia, H. E.; Jiang, L.; Michael, P.

    2012-12-01

    As the designated Federal permanent oceanographic data center in the United States, NOAA's National Oceanographic Data Center (NODC) has been providing scientific stewardship for national and international marine environmental and ecosystem data for over 50 years. NODC is supporting NOAA's Ocean Acidification Program and the science community by providing end-to-end scientific data management of ocean acidification (OA) data, dedicated online data discovery, and user-friendly access to a diverse range of historical and modern OA and other chemical, physical, and biological oceanographic data. This effort is being catalyzed by the NOAA Ocean Acidification Program, but the intended reach is for the broader scientific ocean acidification community. The first three years of the project will be focused on infrastructure building. A complete ocean acidification data content standard is being developed to ensure that a full spectrum of ocean acidification data and metadata can be stored and utilized for optimal data discovery and access in usable data formats. We plan to develop a data access interface capable of allowing users to constrain their search based on real-time and delayed mode measured variables, scientific data quality, their observation types, the temporal coverage, methods, instruments, standards, collecting institutions, and the spatial coverage. In addition, NODC seeks to utilize the existing suite of international standards (including ISO 19115-2 and CF-compliant netCDF) to help our data producers use those standards for their data, and help our data consumers make use of the well-standardized metadata-rich data sets. These tools will be available through our NODC Ocean Acidification Scientific Data Stewardship (OADS) web page at http://www.nodc.noaa.gov/oceanacidification. NODC also has a goal to provide each archived dataset with a unique ID, to ensure a means of providing credit to the data provider. Working with partner institutions, such as the

  7. The end-to-end simulator for the E-ELT HIRES high resolution spectrograph

    NASA Astrophysics Data System (ADS)

    Genoni, M.; Landoni, M.; Riva, M.; Pariani, G.; Mason, E.; Di Marcantonio, P.; Disseau, K.; Di Varano, I.; Gonzalez, O.; Huke, P.; Korhonen, H.; Li Causi, Gianluca

    2017-06-01

    We present the design, architecture and results of the End-to-End simulator model of the high resolution spectrograph HIRES for the European Extremely Large Telescope (E-ELT). This system can be used as a tool to characterize the spectrograph both by engineers and scientists. The model allows to simulate the behavior of photons starting from the scientific object (modeled bearing in mind the main science drivers) to the detector, considering also calibration light sources, and allowing to perform evaluation of the different parameters of the spectrograph design. In this paper, we will detail the architecture of the simulator and the computational model which are strongly characterized by modularity and flexibility that will be crucial in the next generation astronomical observation projects like E-ELT due to of the high complexity and long-time design and development. Finally, we present synthetic images obtained with the current version of the End-to-End simulator based on the E-ELT HIRES requirements (especially high radial velocity accuracy). Once ingested in the Data reduction Software (DRS), they will allow to verify that the instrument design can achieve the radial velocity accuracy needed by the HIRES science cases.

  8. An End-To-End Near-Earth Asteroid Resource Exploitation Plan

    NASA Astrophysics Data System (ADS)

    Reed, K. L.

    2000-01-01

    The possible end result of the utilization of raw materials garnered from near-Earth asteroids (NEAs) has been well documented if often a bit fanciful. Very few have put forward an end-to-end plan from prospecting to mine closure for any specific asteroid or for any particular asteroid resource. There are many aspects to planning for the mining of raw materials from asteroids that have never been encountered in terrestrial resource exploitation due to the dispersed nature of the asteroids. As an example from petroleum exploration, if a dry hole is drilled in a large geologic setting indicative of petroleum deposits, one only need pack the drill rig up and move it to a new spot. In asteroid exploitation, the problem of "moving to a new spot" is complicated, as the "new spot" is moving constantly and may be many millions of kilometers distant at great cost in time and rocket fuel. This paper will outline a relatively low-risk, probable high-return, end-to-end plan for the exploitation and utilization of asteroid raw materials. All aspects of exploration and mining will attempt to be addressed, from prospecting, exploration, and evaluation of possible resources to initialization, industrialization, and closure of the mine. It will attempt to plan for the acquisition of not just the needed scientific knowledge, but also to plan for acquisition of the engineering and geotechnical knowledge needed for effective mining of a small planetary object.

  9. An End-To-End Test of A Simulated Nuclear Electric Propulsion System

    NASA Technical Reports Server (NTRS)

    VanDyke, Melissa; Hrbud, Ivana; Goddfellow, Keith; Rodgers, Stephen L. (Technical Monitor)

    2002-01-01

    The Safe Affordable Fission Engine (SAFE) test series addresses Phase I Space Fission Systems issues in it particular non-nuclear testing and system integration issues leading to the testing and non-nuclear demonstration of a 400-kW fully integrated flight unit. The first part of the SAFE 30 test series demonstrated operation of the simulated nuclear core and heat pipe system. Experimental data acquired in a number of different test scenarios will validate existing computational models, demonstrated system flexibility (fast start-ups, multiple start-ups/shut downs), simulate predictable failure modes and operating environments. The objective of the second part is to demonstrate an integrated propulsion system consisting of a core, conversion system and a thruster where the system converts thermal heat into jet power. This end-to-end system demonstration sets a precedent for ground testing of nuclear electric propulsion systems. The paper describes the SAFE 30 end-to-end system demonstration and its subsystems.

  10. A multicentre 'end to end' dosimetry audit for cervix HDR brachytherapy treatment.

    PubMed

    Palmer, Antony L; Diez, Patricia; Gandon, Laura; Wynn-Jones, Andrea; Bownes, Peter; Lee, Chris; Aird, Edwin; Bidmead, Margaret; Lowe, Gerry; Bradley, David; Nisbet, Andrew

    2015-02-01

    To undertake the first multicentre fully 'end to end' dosimetry audit for HDR cervix brachytherapy, comparing planned and delivered dose distributions around clinical treatment applicators, with review of local procedures. A film-dosimetry audit was performed at 46 centres, including imaging, applicator reconstruction, treatment planning and delivery. Film dose maps were calculated using triple-channel dosimetry and compared to RTDose data from treatment planning systems. Deviations between plan and measurement were quantified at prescription Point A and using gamma analysis. Local procedures were also discussed. The mean difference between planned and measured dose at Point A was -0.6% for plastic applicators and -3.0% for metal applicators, at standard uncertainty 3.0% (k=1). Isodose distributions agreed within 1mm over a dose range 2-16Gy. Mean gamma passing rates exceeded 97% for plastic and metal applicators at 3% (local) 2mm criteria. Two errors were found: one dose normalisation error and one applicator library misaligned with the imaged applicator. Suggestions for quality improvement were also made. The concept of 'end to end' dosimetry audit for HDR brachytherapy has been successfully implemented in a multicentre environment, providing evidence that a high level of accuracy in brachytherapy dosimetry can be achieved. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  11. End-to-End Rate-Distortion Optimized MD Mode Selection for Multiple Description Video Coding

    NASA Astrophysics Data System (ADS)

    Heng, Brian A.; Apostolopoulos, John G.; Lim, Jae S.

    2006-12-01

    Multiple description (MD) video coding can be used to reduce the detrimental effects caused by transmission over lossy packet networks. A number of approaches have been proposed for MD coding, where each provides a different tradeoff between compression efficiency and error resilience. How effectively each method achieves this tradeoff depends on the network conditions as well as on the characteristics of the video itself. This paper proposes an adaptive MD coding approach which adapts to these conditions through the use of adaptive MD mode selection. The encoder in this system is able to accurately estimate the expected end-to-end distortion, accounting for both compression and packet loss-induced distortions, as well as for the bursty nature of channel losses and the effective use of multiple transmission paths. With this model of the expected end-to-end distortion, the encoder selects between MD coding modes in a rate-distortion (R-D) optimized manner to most effectively tradeoff compression efficiency for error resilience. We show how this approach adapts to both the local characteristics of the video and network conditions and demonstrates the resulting gains in performance using an H.264-based adaptive MD video coder.

  12. Micro-Precision Interferometer Testbed: end-to-end system integration of control structure interaction technologies

    NASA Astrophysics Data System (ADS)

    Neat, Gregory W.; Sword, Lee F.; Hines, Braden E.; Calvet, Robert J.

    1993-09-01

    This paper describes the overall design and planned phased delivery of the ground-based Micro-Precision Interferometer (MPI) Testbed. The testbed is a half scale replica of a future space-based interferometer containing all the spacecraft subsystems necessary to perform an astrometric measurement. Appropriate sized reaction wheels will regulate the testbed attitude as well as provide a flight-like disturbance source. The optical system will consist of two complete Michelson interferometers. Successful interferometric measurements require controlling the positional stabilities of these optical elements to the nanometer level. The primary objective of the testbed is to perform a system integration of Control Structure Interaction (CSI) technologies necessary to demonstrate the end-to-end operation of a space- based interferometer, ultimately proving to flight mission planners that the necessary control technology exists to meet the challenging requirements of future space-based interferometry missions. These technologies form a multi-layered vibration attenuation architecture to achieve the necessary quiet environment. This three layered methodology blends disturbance isolation, structural quieting and active optical control techniques. The paper describes all the testbed subsystems in this end-to-end ground-based system as well as the present capabilities of the evolving testbed.

  13. Cyberinfrastructure to support Real-time, End-to-End, High Resolution, Localized Forecasting

    NASA Astrophysics Data System (ADS)

    Ramamurthy, M. K.; Lindholm, D.; Baltzer, T.; Domenico, B.

    2004-12-01

    From natural disasters such as flooding and forest fires to man-made disasters such as toxic gas releases, the impact of weather-influenced severe events on society can be profound. Understanding, predicting, and mitigating such local, mesoscale events calls for a cyberinfrastructure to integrate multidisciplinary data, tools, and services as well as the capability to generate and use high resolution data (such as wind and precipitation) from localized models. The need for such end to end systems -- including data collection, distribution, integration, assimilation, regionalized mesoscale modeling, analysis, and visualization -- has been realized to some extent in many academic and quasi-operational environments, especially for atmospheric sciences data. However, many challenges still remain in the integration and synthesis of data from multiple sources and the development of interoperable data systems and services across those disciplines. Over the years, the Unidata Program Center has developed several tools that have either directly or indirectly facilitated these local modeling activities. For example, the community is using Unidata technologies such as the Internet Data Distribution (IDD) system, Local Data Manger (LDM), decoders, netCDF libraries, Thematic Realtime Environmental Distributed Data Services (THREDDS), and the Integrated Data Viewer (IDV) in their real-time prediction efforts. In essence, these technologies for data reception and processing, local and remote access, cataloging, and analysis and visualization coupled with technologies from others in the community are becoming the foundation of a cyberinfrastructure to support an end-to-end regional forecasting system. To build on these capabilities, the Unidata Program Center is pleased to be a significant contributor to the Linked Environments for Atmospheric Discovery (LEAD) project, a NSF-funded multi-institutional large Information Technology Research effort. The goal of LEAD is to create an

  14. End-to-end performance analysis using engineering confidence models and a ground processor prototype

    NASA Astrophysics Data System (ADS)

    Kruse, Klaus-Werner; Sauer, Maximilian; Jäger, Thomas; Herzog, Alexandra; Schmitt, Michael; Huchler, Markus; Wallace, Kotska; Eisinger, Michael; Heliere, Arnaud; Lefebvre, Alain; Maher, Mat; Chang, Mark; Phillips, Tracy; Knight, Steve; de Goeij, Bryan T. G.; van der Knaap, Frits; Van't Hof, Adriaan

    2015-10-01

    The European Space Agency (ESA) and the Japan Aerospace Exploration Agency (JAXA) are co-operating to develop the EarthCARE satellite mission with the fundamental objective of improving the understanding of the processes involving clouds, aerosols and radiation in the Earth's atmosphere. The EarthCARE Multispectral Imager (MSI) is relatively compact for a space borne imager. As a consequence, the immediate point-spread function (PSF) of the instrument will be mainly determined by the diffraction caused by the relatively small optical aperture. In order to still achieve a high contrast image, de-convolution processing is applied to remove the impact of diffraction on the PSF. A Lucy-Richardson algorithm has been chosen for this purpose. This paper will describe the system setup and the necessary data pre-processing and post-processing steps applied in order to compare the end-to-end image quality with the L1b performance required by the science community.

  15. End-to-End QoS for Differentiated Services and ATM Internetworking

    NASA Technical Reports Server (NTRS)

    Su, Hongjun; Atiquzzaman, Mohammed

    2001-01-01

    The Internet was initially design for non real-time data communications and hence does not provide any Quality of Service (QoS). The next generation Internet will be characterized by high speed and QoS guarantee. The aim of this paper is to develop a prioritized early packet discard (PEPD) scheme for ATM switches to provide service differentiation and QoS guarantee to end applications running over next generation Internet. The proposed PEPD scheme differs from previous schemes by taking into account the priority of packets generated from different application. We develop a Markov chain model for the proposed scheme and verify the model with simulation. Numerical results show that the results from the model and computer simulation are in close agreement. Our PEPD scheme provides service differentiation to the end-to-end applications.

  16. End-to-end simulations of the E-ELT/METIS coronagraphs

    NASA Astrophysics Data System (ADS)

    Carlomagno, Brunella; Absil, Olivier; Kenworthy, Matthew; Ruane, Garreth; Keller, Christoph U.; Otten, Gilles; Feldt, Markus; Hippler, Stefan; Huby, Elsa; Mawet, Dimitri; Delacroix, Christian; Surdej, Jean; Habraken, Serge; Forsberg, Pontus; Karlsson, Mikael; Vargas Catalan, Ernesto; Brandl, Bernhard R.

    2016-07-01

    The direct detection of low-mass planets in the habitable zone of nearby stars is an important science case for future E-ELT instruments such as the mid-infrared imager and spectrograph METIS, which features vortex phase masks and apodizing phase plates (APP) in its baseline design. In this work, we present end-to-end performance simulations, using Fourier propagation, of several METIS coronagraphic modes, including focal-plane vortex phase masks and pupil-plane apodizing phase plates, for the centrally obscured, segmented E-ELT pupil. The atmosphere and the AO contributions are taken into account. Hybrid coronagraphs combining the advantages of vortex phase masks and APPs are considered to improve the METIS coronagraphic performance.

  17. The Kepler End-to-End Data Pipeline: From Photons to Far Away Worlds

    NASA Technical Reports Server (NTRS)

    Cooke, Brian; Thompson, Richard; Standley, Shaun

    2012-01-01

    Launched by NASA on 6 March 2009, the Kepler Mission has been observing more than 100,000 targets in a single patch of sky between the constellations Cygnus and Lyra almost continuously for the last two years looking for planetary systems using the transit method. As of October 2011, the Kepler spacecraft has collected and returned to Earth just over 290 GB of data, identifying 1235 planet candidates with 25 of these candidates confirmed as planets via ground observation. Extracting the telltale signature of a planetary system from stellar photometry where valid signal transients can be small as a 40 ppm is a difficult and exacting task. The end-to end processing of determining planetary candidates from noisy, raw photometric measurements is discussed.

  18. End-to-end automated microfluidic platform for synthetic biology: from design to functional analysis

    DOE PAGES

    Linshiz, Gregory; Jensen, Erik; Stawski, Nina; ...

    2016-02-02

    Synthetic biology aims to engineer biological systems for desired behaviors. The construction of these systems can be complex, often requiring genetic reprogramming, extensive de novo DNA synthesis, and functional screening. Here, we present a programmable, multipurpose microfluidic platform and associated software and apply the platform to major steps of the synthetic biology research cycle: design, construction, testing, and analysis. We show the platform’s capabilities for multiple automated DNA assembly methods, including a new method for Isothermal Hierarchical DNA Construction, and for Escherichia coli and Saccharomyces cerevisiae transformation. The platform enables the automated control of cellular growth, gene expression induction, andmore » proteogenic and metabolic output analysis. Finally, taken together, we demonstrate the microfluidic platform’s potential to provide end-to-end solutions for synthetic biology research, from design to functional analysis.« less

  19. End-to-End Assessment of a Large Aperture Segmented Ultraviolet Optical Infrared (UVOIR) Telescope Architecture

    NASA Technical Reports Server (NTRS)

    Feinberg, Lee; Rioux, Norman; Bolcar, Matthew; Liu, Alice; Guyon, Oliver; Stark, Chris; Arenberg, Jon

    2016-01-01

    Key challenges of a future large aperture, segmented Ultraviolet Optical Infrared (UVOIR) Telescope capable of performing a spectroscopic survey of hundreds of Exoplanets will be sufficient stability to achieve 10^-10 contrast measurements and sufficient throughput and sensitivity for high yield Exo-Earth spectroscopic detection. Our team has collectively assessed an optimized end to end architecture including a high throughput coronagraph capable of working with a segmented telescope, a cost-effective and heritage based stable segmented telescope, a control architecture that minimizes the amount of new technologies, and an Exo-Earth yield assessment to evaluate potential performance. These efforts are combined through integrated modeling, coronagraph evaluations, and Exo-Earth yield calculations to assess the potential performance of the selected architecture. In addition, we discusses the scalability of this architecture to larger apertures and the technological tall poles to enabling it.

  20. Orion MPCV GN and C End-to-End Phasing Tests

    NASA Technical Reports Server (NTRS)

    Neumann, Brian C.

    2013-01-01

    End-to-end integration tests are critical risk reduction efforts for any complex vehicle. Phasing tests are an end-to-end integrated test that validates system directional phasing (polarity) from sensor measurement through software algorithms to end effector response. Phasing tests are typically performed on a fully integrated and assembled flight vehicle where sensors are stimulated by moving the vehicle and the effectors are observed for proper polarity. Orion Multi-Purpose Crew Vehicle (MPCV) Pad Abort 1 (PA-1) Phasing Test was conducted from inertial measurement to Launch Abort System (LAS). Orion Exploration Flight Test 1 (EFT-1) has two end-to-end phasing tests planned. The first test from inertial measurement to Crew Module (CM) reaction control system thrusters uses navigation and flight control system software algorithms to process commands. The second test from inertial measurement to CM S-Band Phased Array Antenna (PAA) uses navigation and communication system software algorithms to process commands. Future Orion flights include Ascent Abort Flight Test 2 (AA-2) and Exploration Mission 1 (EM-1). These flights will include additional or updated sensors, software algorithms and effectors. This paper will explore the implementation of end-to-end phasing tests on a flight vehicle which has many constraints, trade-offs and compromises. Orion PA-1 Phasing Test was conducted at White Sands Missile Range (WSMR) from March 4-6, 2010. This test decreased the risk of mission failure by demonstrating proper flight control system polarity. Demonstration was achieved by stimulating the primary navigation sensor, processing sensor data to commands and viewing propulsion response. PA-1 primary navigation sensor was a Space Integrated Inertial Navigation System (INS) and Global Positioning System (GPS) (SIGI) which has onboard processing, INS (3 accelerometers and 3 rate gyros) and no GPS receiver. SIGI data was processed by GN&C software into thrust magnitude and

  1. End-to-end interoperability and workflows from building architecture design to one or more simulations

    DOEpatents

    Chao, Tian-Jy; Kim, Younghun

    2015-02-10

    An end-to-end interoperability and workflows from building architecture design to one or more simulations, in one aspect, may comprise establishing a BIM enablement platform architecture. A data model defines data entities and entity relationships for enabling the interoperability and workflows. A data definition language may be implemented that defines and creates a table schema of a database associated with the data model. Data management services and/or application programming interfaces may be implemented for interacting with the data model. Web services may also be provided for interacting with the data model via the Web. A user interface may be implemented that communicates with users and uses the BIM enablement platform architecture, the data model, the data definition language, data management services and application programming interfaces to provide functions to the users to perform work related to building information management.

  2. An end-to-end secure patient information access card system.

    PubMed

    Alkhateeb, A; Singer, H; Yakami, M; Takahashi, T

    2000-03-01

    The rapid development of the Internet and the increasing interest in Internet-based solutions has promoted the idea of creating Internet-based health information applications. This will force a change in the role of IC cards in healthcare card systems from a data carrier to an access key medium. At the Medical Informatics Department of Kyoto University Hospital we are developing a smart card patient information project where patient databases are accessed via the Internet. Strong end-to-end data encryption is performed via Secure Socket Layers, transparent to transmit patient information. The smart card is playing the crucial role of access key to the database: user authentication is performed internally without ever revealing the actual key. For easy acceptance by healthcare professionals, the user interface is integrated as a plug-in for two familiar Web browsers, Netscape Navigator and MS Internet Explorer.

  3. Enhancing End-to-End Performance of Information Services Over Ka-Band Global Satellite Networks

    NASA Technical Reports Server (NTRS)

    Bhasin, Kul B.; Glover, Daniel R.; Ivancic, William D.; vonDeak, Thomas C.

    1997-01-01

    The Internet has been growing at a rapid rate as the key medium to provide information services such as e-mail, WWW and multimedia etc., however its global reach is limited. Ka-band communication satellite networks are being developed to increase the accessibility of information services via the Internet at global scale. There is need to assess satellite networks in their ability to provide these services and interconnect seamlessly with existing and proposed terrestrial telecommunication networks. In this paper the significant issues and requirements in providing end-to-end high performance for the delivery of information services over satellite networks based on various layers in the OSI reference model are identified. Key experiments have been performed to evaluate the performance of digital video and Internet over satellite-like testbeds. The results of the early developments in ATM and TCP protocols over satellite networks are summarized.

  4. End-to-end test of the electron-proton spectrometer

    NASA Technical Reports Server (NTRS)

    Cash, B. L.

    1972-01-01

    A series of end-to-end tests were performed to demonstrate the proper functioning of the complete Electron-Proton Spectrometer (EPS). The purpose of the tests was to provide experimental verification of the design and to provide a complete functional performance check of the instrument from the excitation of the sensors to and including the data processor and equipment test set. Each of the channels of the EPS was exposed to a calibrated beam of energetic particles, and counts were accumulated for a predetermined period of time for each of several energies. The counts were related to the known flux of particles to give a monodirectional response function for each channel. The measured response function of the test unit was compared to the response function determined for the calibration sensors from the data taken from the calibration program.

  5. End-to-End Network Simulation Using a Site-Specific Radio Wave Propagation Model

    SciTech Connect

    Djouadi, Seddik M; Kuruganti, Phani Teja; Nutaro, James J

    2013-01-01

    The performance of systems that rely on a wireless network depends on the propagation environment in which that network operates. To predict how these systems and their supporting networks will perform, simulations must take into consideration the propagation environment and how this effects the performance of the wireless network. Network simulators typically use empirical models of the propagation environment. However, these models are not intended for, and cannot be used, to predict a wireless system will perform in a specific location, e.g., in the center of a particular city or the interior of a specific manufacturing facility. In this paper, we demonstrate how a site-specific propagation model and the NS3 simulator can be used to predict the end-to-end performance of a wireless network.

  6. End-to-end performance measurement of Internet based medical applications.

    PubMed Central

    Dev, P.; Harris, D.; Gutierrez, D.; Shah, A.; Senger, S.

    2002-01-01

    We present a method to obtain an end-to-end characterization of the performance of an application over a network. This method is not dependent on any specific application or type of network. The method requires characterization of network parameters, such as latency and packet loss, between the expected server or client endpoints, as well as characterization of the application's constraints on these parameters. A subjective metric is presented that integrates these characterizations and that operates over a wide range of applications and networks. We believe that this method may be of wide applicability as research and educational applications increasingly make use of computation and data servers that are distributed over the Internet. PMID:12463816

  7. Sieving of H2 and D2 Through End-to-End Nanotubes

    NASA Astrophysics Data System (ADS)

    Devagnik, Dasgupta; Debra, J. Searles; Lamberto, Rondoni; Stefano, Bernardi

    2014-10-01

    We study the quantum molecular sieving of H2 and D2 through two nanotubes placed end-to-end. An analytic treatment, assuming that the particles have classical motion along the axis of the nanotube and are confined in a potential well in the radial direction, is considered. Using this idealistic model, and under certain conditions, it is found that this device can act as a complete sieve, allowing chemically pure deuterium to be isolated from an isotope mixture. We also consider a more realistic model of two carbon nanotubes and carry out molecular dynamics simulations using a Feynman—Hibbs potential to model the quantum effects on the dynamics of H2 and D2. Sieving is also observed in this case, but is caused by a different process.

  8. End-to-end commissioning demonstration of the James Webb Space Telescope

    NASA Astrophysics Data System (ADS)

    Acton, D. Scott; Towell, Timothy; Schwenker, John; Shields, Duncan; Sabatke, Erin; Contos, Adam R.; Hansen, Karl; Shi, Fang; Dean, Bruce; Smith, Scott

    2007-09-01

    The one-meter Testbed Telescope (TBT) has been developed at Ball Aerospace to facilitate the design and implementation of the wavefront sensing and control (WFSC) capabilities of the James Webb Space Telescope (JWST). We have recently conducted an "end-to-end" demonstration of the flight commissioning process on the TBT. This demonstration started with the Primary Mirror (PM) segments and the Secondary Mirror (SM) in random positions, traceable to the worst-case flight deployment conditions. The commissioning process detected and corrected the deployment errors, resulting in diffraction-limited performance across the entire science FOV. This paper will describe the commissioning demonstration and the WFSC algorithms used at each step in the process.

  9. End-to-end automated microfluidic platform for synthetic biology: from design to functional analysis

    SciTech Connect

    Linshiz, Gregory; Jensen, Erik; Stawski, Nina; Bi, Changhao; Elsbree, Nick; Jiao, Hong; Kim, Jungkyu; Mathies, Richard; Keasling, Jay D.; Hillson, Nathan J.

    2016-02-02

    Synthetic biology aims to engineer biological systems for desired behaviors. The construction of these systems can be complex, often requiring genetic reprogramming, extensive de novo DNA synthesis, and functional screening. Here, we present a programmable, multipurpose microfluidic platform and associated software and apply the platform to major steps of the synthetic biology research cycle: design, construction, testing, and analysis. We show the platform’s capabilities for multiple automated DNA assembly methods, including a new method for Isothermal Hierarchical DNA Construction, and for Escherichia coli and Saccharomyces cerevisiae transformation. The platform enables the automated control of cellular growth, gene expression induction, and proteogenic and metabolic output analysis. Finally, taken together, we demonstrate the microfluidic platform’s potential to provide end-to-end solutions for synthetic biology research, from design to functional analysis.

  10. The Consolidation of the End-to-End Avionics Systems Testbench

    NASA Astrophysics Data System (ADS)

    Wijnands, Quirien; Torelli, Felice; Blommestijn, Robert; Kranz, Stephan; Koster, Jean-Paul

    2014-08-01

    Over the past years, the Avionics System Test Bench (ATB) has been used to support the demonstration and validation of upcoming space avionics related standards and technologies in a representative environment. Next to this another main use-case of the facility has been to support projects in their needs of assessing particular technology related issues. In doing so, it was necessary to add activity- and project specifics to different configurations of the ATB, leading to a proliferation of facilities and technologies. In some cases however the results and lessons-learned from these efforts and activities were considered valuable to the ATB-concept in general and therefore needed preservation in the ATB mainstream for future reuse. Currently activities are ongoing to consolidate the End-To-End Avionics Systems TestBench (E2E-ATB). In this paper the resulting details of these activities are described as enhancements and improvements per ATB configuration.

  11. End-to-End Beam Simulations for the New Muon G-2 Experiment at Fermilab

    SciTech Connect

    Korostelev, Maxim; Bailey, Ian; Herrod, Alexander; Morgan, James; Morse, William; Stratakis, Diktys; Tishchenko, Vladimir; Wolski, Andrzej

    2016-06-01

    The aim of the new muon g-2 experiment at Fermilab is to measure the anomalous magnetic moment of the muon with an unprecedented uncertainty of 140 ppb. A beam of positive muons required for the experiment is created by pion decay. Detailed studies of the beam dynamics and spin polarization of the muons are important to predict systematic uncertainties in the experiment. In this paper, we present the results of beam simulations and spin tracking from the pion production target to the muon storage ring. The end-to-end beam simulations are developed in Bmad and include the processes of particle decay, collimation (with accurate representation of all apertures) and spin tracking.

  12. Swift XRT effective area measured at the Panter end-to-end tests

    NASA Astrophysics Data System (ADS)

    Tagliaferri, Gianpiero; Moretti, Alberto; Campana, Sergio; Abbey, Anthony F.; Ambrosi, Richard M.; Angelini, Lorella; Beardmore, Andrew P.; Bräuninger, Heinrich W.; Burkert, Wolfgang; Burrows, David N.; Capalbi, Milvia; Chincarini, Guido; Citterio, Oberto; Cusumano, Giancarlo; Freyberg, Michael J.; Giommi, Paolo; Hartner, Gisela D.; Hill, Joanne E.; Mori, Koji; Morris, Dave C.; Mukerjee, Kallol; Nousek, John A.; Osborne, Julian P.; Short, Alexander D. T.; Tamburelli, F.; Watson, D. J.; Wells, Alan A.

    2004-02-01

    The Swift X-ray Telescope (XRT) is designed to make astrometric, spectroscopic and photometric observations of the X-ray emission from Gamma-ray bursts and their afterglows, in the energy band 0.2 - 10 keV. Here we report first results of the analysis of Swift XRT effective area at five different energies as measured during the end-to-end calibration campaign at the Panter X-ray beam line facility. The analysis comprises the study of the effective area both on-axis and off-axis for different event grade selection. We compare the laboratory results with the expectations and show that the measured effective area meets the mission scientific requirements.

  13. End-to-end automated microfluidic platform for synthetic biology: from design to functional analysis

    DOE PAGES

    Linshiz, Gregory; Jensen, Erik; Stawski, Nina; ...

    2016-02-02

    Synthetic biology aims to engineer biological systems for desired behaviors. The construction of these systems can be complex, often requiring genetic reprogramming, extensive de novo DNA synthesis, and functional screening. Here, we present a programmable, multipurpose microfluidic platform and associated software and apply the platform to major steps of the synthetic biology research cycle: design, construction, testing, and analysis. We show the platform’s capabilities for multiple automated DNA assembly methods, including a new method for Isothermal Hierarchical DNA Construction, and for Escherichia coli and Saccharomyces cerevisiae transformation. The platform enables the automated control of cellular growth, gene expression induction, andmore » proteogenic and metabolic output analysis. Finally, taken together, we demonstrate the microfluidic platform’s potential to provide end-to-end solutions for synthetic biology research, from design to functional analysis.« less

  14. Data analysis pipeline for EChO end-to-end simulations

    NASA Astrophysics Data System (ADS)

    Waldmann, Ingo P.; Pascale, E.

    2015-12-01

    Atmospheric spectroscopy of extrasolar planets is an intricate business. Atmospheric signatures typically require a photometric precision of 1×10-4 in flux over several hours. Such precision demands high instrument stability as well as an understanding of stellar variability and an optimal data reduction and removal of systematic noise. In the context of the EChO mission concept, we here discuss the data reduction and analysis pipeline developed for the EChO end-to-end simulator EChOSim. We present and discuss the step by step procedures required in order to obtain the final exoplanetary spectrum from the EChOSim `raw data' using a simulated observation of the secondary eclipse of the hot-Neptune 55 Cnc e.

  15. End-to-end system test for solid-state microdosemeters.

    PubMed

    Pisacane, V L; Dolecek, Q E; Malak, H; Dicello, J F

    2010-08-01

    The gold standard in microdosemeters has been the tissue equivalent proportional counter (TEPC) that utilises a gas cavity. An alternative is the solid-state microdosemeter that replaces the gas with a condensed phase (silicon) detector with microscopic sensitive volumes. Calibrations of gas and solid-state microdosemeters are generally carried out using radiation sources built into the detector that impose restrictions on their handling, transportation and licensing in accordance with the regulations from international, national and local nuclear regulatory bodies. Here a novel method is presented for carrying out a calibration and end-to-end system test of a microdosemeter using low-energy photons as the initiating energy source, thus obviating the need for a regulated ionising radiation source. This technique may be utilised to calibrate both a solid-state microdosemeter and, with modification, a TEPC with the higher average ionisation energy of a gas.

  16. End-to-end assessment of a large aperture segmented ultraviolet optical infrared (UVOIR) telescope architecture

    NASA Astrophysics Data System (ADS)

    Feinberg, Lee; Rioux, Norman; Bolcar, Matthew; Liu, Alice; Guyon, Olivier; Stark, Chris; Arenberg, Jon

    2016-07-01

    Key challenges of a future large aperture, segmented Ultraviolet Optical Infrared (UVOIR) Telescope capable of performing a spectroscopic survey of hundreds of Exoplanets will be sufficient stability to achieve 10^-10 contrast measurements and sufficient throughput and sensitivity for high yield exo-earth spectroscopic detection. Our team has collectively assessed an optimized end to end architecture including a high throughput coronagraph capable of working with a segmented telescope, a cost-effective and heritage based stable segmented telescope, a control architecture that minimizes the amount of new technologies, and an exo-earth yield assessment to evaluate potential performance. These efforts are combined through integrated modeling, coronagraph evaluations, and exo-earth yield calculations to assess the potential performance of the selected architecture. In addition, we discusses the scalability of this architecture to larger apertures and the technological tall poles to enabling these missions.

  17. The Kepler End-to-End Data Pipeline: From Photons to Far Away Worlds

    NASA Technical Reports Server (NTRS)

    Cooke, Brian; Thompson, Richard; Standley, Shaun

    2012-01-01

    Launched by NASA on 6 March 2009, the Kepler Mission has been observing more than 100,000 targets in a single patch of sky between the constellations Cygnus and Lyra almost continuously for the last two years looking for planetary systems using the transit method. As of October 2011, the Kepler spacecraft has collected and returned to Earth just over 290 GB of data, identifying 1235 planet candidates with 25 of these candidates confirmed as planets via ground observation. Extracting the telltale signature of a planetary system from stellar photometry where valid signal transients can be small as a 40 ppm is a difficult and exacting task. The end-to end processing of determining planetary candidates from noisy, raw photometric measurements is discussed.

  18. End-to-end automated microfluidic platform for synthetic biology: from design to functional analysis

    SciTech Connect

    Linshiz, Gregory; Jensen, Erik; Stawski, Nina; Bi, Changhao; Elsbree, Nick; Jiao, Hong; Kim, Jungkyu; Mathies, Richard; Keasling, Jay D.; Hillson, Nathan J.

    2016-02-02

    Synthetic biology aims to engineer biological systems for desired behaviors. The construction of these systems can be complex, often requiring genetic reprogramming, extensive de novo DNA synthesis, and functional screening. Here, we present a programmable, multipurpose microfluidic platform and associated software and apply the platform to major steps of the synthetic biology research cycle: design, construction, testing, and analysis. We show the platform’s capabilities for multiple automated DNA assembly methods, including a new method for Isothermal Hierarchical DNA Construction, and for Escherichia coli and Saccharomyces cerevisiae transformation. The platform enables the automated control of cellular growth, gene expression induction, and proteogenic and metabolic output analysis. Finally, taken together, we demonstrate the microfluidic platform’s potential to provide end-to-end solutions for synthetic biology research, from design to functional analysis.

  19. End-to-end communication test on variable length packet structures utilizing AOS testbed

    NASA Technical Reports Server (NTRS)

    Miller, Warner H.; Sank, V.; Fong, Wai; Miko, J.; Powers, M.; Folk, John; Conaway, B.; Michael, K.; Yeh, Pen-Shu

    1994-01-01

    This paper describes a communication test, which successfully demonstrated the transfer of losslessly compressed images in an end-to-end system. These compressed images were first formatted into variable length Consultative Committee for Space Data Systems (CCSDS) packets in the Advanced Orbiting System Testbed (AOST). The CCSDS data Structures were transferred from the AOST to the Radio Frequency Simulations Operations Center (RFSOC), via a fiber optic link, where data was then transmitted through the Tracking and Data Relay Satellite System (TDRSS). The received data acquired at the White Sands Complex (WSC) was transferred back to the AOST where the data was captured and decompressed back to the original images. This paper describes the compression algorithm, the AOST configuration, key flight components, data formats, and the communication link characteristics and test results.

  20. Advances in POST2 End-to-End Descent and Landing Simulation for the ALHAT Project

    NASA Technical Reports Server (NTRS)

    Davis, Jody L.; Striepe, Scott A.; Maddock, Robert W.; Hines, Glenn D.; Paschall, Stephen, II; Cohanim, Babak E.; Fill, Thomas; Johnson, Michael C.; Bishop, Robert H.; DeMars, Kyle J.; hide

    2008-01-01

    Program to Optimize Simulated Trajectories II (POST2) is used as a basis for an end-to-end descent and landing trajectory simulation that is essential in determining design and integration capability and system performance of the lunar descent and landing system and environment models for the Autonomous Landing and Hazard Avoidance Technology (ALHAT) project. The POST2 simulation provides a six degree-of-freedom capability necessary to test, design and operate a descent and landing system for successful lunar landing. This paper presents advances in the development and model-implementation of the POST2 simulation, as well as preliminary system performance analysis, used for the testing and evaluation of ALHAT project system models.

  1. Development of a Dynamic, End-to-End Free Piston Stirling Convertor Model

    NASA Technical Reports Server (NTRS)

    Regan, Timothy F.; Gerber, Scott S.; Roth, Mary Ellen

    2004-01-01

    A dynamic model for a free-piston Stirling convertor is being developed at the NASA Glenn Research Center. The model is an end-to-end system model that includes the cycle thermodynamics, the dynamics, and electrical aspects of the system. The subsystems of interest are the heat source, the springs, the moving masses, the linear alternator, the controller, and the end-user load. The envisioned use of the model will be in evaluating how changes in a subsystem could affect the operation of the convertor. The model under development will speed the evaluation of improvements to a subsystem and aid in determining areas in which most significant improvements may be found. One of the first uses of the end-toend model will be in the development of controller architectures. Another related area is in evaluating changes to details in the linear alternator.

  2. Evaluation of NASA's end-to-end data systems using DSDS+

    NASA Technical Reports Server (NTRS)

    Rouff, Christopher; Davenport, William; Message, Philip

    1994-01-01

    The Data Systems Dynamic Simulator (DSDS+) is a software tool being developed by the authors to evaluate candidate architectures for NASA's end-to-end data systems. Via modeling and simulation, we are able to quickly predict the performance characteristics of each architecture, to evaluate 'what-if' scenarios, and to perform sensitivity analyses. As such, we are using modeling and simulation to help NASA select the optimal system configuration, and to quantify the performance characteristics of this system prior to its delivery. This paper is divided into the following six sections: (1) The role of modeling and simulation in the systems engineering process. In this section, we briefly describe the different types of results obtained by modeling each phase of the systems engineering life cycle, from concept definition through operations and maintenance; (2) Recent applications of DSDS+. In this section, we describe ongoing applications of DSDS+ in support of the Earth Observing System (EOS), and we present some of the simulation results generated of candidate system designs. So far, we have modeled individual EOS subsystems (e.g. the Solid State Recorders used onboard the spacecraft), and we have also developed an integrated model of the EOS end-to-end data processing and data communications systems (from the payloads onboard to the principle investigator facilities on the ground); (3) Overview of DSDS+. In this section we define what a discrete-event model is, and how it works. The discussion is presented relative to the DSDS+ simulation tool that we have developed, including it's run-time optimization algorithms that enables DSDS+ to execute substantially faster than comparable discrete-event simulation tools; (4) Summary. In this section, we summarize our findings and 'lessons learned' during the development and application of DSDS+ to model NASA's data systems; (5) Further Information; and (6) Acknowledgements.

  3. Somatic cells efficiently join unrelated DNA segments end-to-end.

    PubMed Central

    Wilson, J H; Berget, P B; Pipas, J M

    1982-01-01

    Molecular substrates for probing nonhomologous recombination in somatic cells were constructed by inserting pBR322 sequences at selected sites on the simian virus 40 (SV40) genome. The chimeric products are too large to be packaged into an SV40 capsid. Therefore, production of viable progeny requires that most of the pBR322 sequences be deleted without altering any SV40 sequences that are essential for lytic infection. As judged by plaque assay, these recombination events occur at readily detectable frequencies after transfection into CV1 monkey kidney cells. Depending on the site of pBR322 insertion, the infectivities of the full-length circular or linear chimeras ranged from 0.02 to 2% of the infectivity of linear wild-type SV40 DNA. Nucleotide sequence analysis of several recombinant progeny revealed three distinct classes of recombination junction and indicated that the causative recombination events were minimally dependent on sequence homology. Potential mechanisms involving recombination at internal sites or at ends were distinguished by measuring the infectivity of chimeric molecules from which various lengths of pBR322 had been removed. These data support end-to-end joining as the primary mechanism by which DNA segments recombine nonhomologously in somatic cells. This end joining appears to be very efficient, since SV40 genomes with complementary single-stranded tails or with short non-complementary pBR322 tails were comparably infectious. Overall, this study indicates that mammalian somatic cells are quite efficient at the willy-nilly end-to-end joining of unrelated DNA segments. Images PMID:6294502

  4. The end-to-end testbed of the optical metrology system on-board LISA Pathfinder

    NASA Astrophysics Data System (ADS)

    Steier, F.; Guzmán Cervantes, F.; García Marín, A. F.; Gerardi, D.; Heinzel, G.; Danzmann, K.

    2009-05-01

    LISA Pathfinder is a technology demonstration mission for the Laser Interferometer Space Antenna (LISA). The main experiment on-board LISA Pathfinder is the so-called LISA Technology Package (LTP) which has the aim to measure the differential acceleration between two free-falling test masses with an accuracy of 3 × 10-14 ms-2 Hz-1/2 between 1 mHz and 30 mHz. This measurement is performed interferometrically by the optical metrology system (OMS) on-board LISA Pathfinder. In this paper, we present the development of an experimental end-to-end testbed of the entire OMS. It includes the interferometer and its sub-units, the interferometer backend which is a phasemeter and the processing of the phasemeter output data. Furthermore, three-axes piezo-actuated mirrors are used instead of the free-falling test masses for the characterization of the dynamic behaviour of the system and some parts of the drag-free and attitude control system (DFACS) which controls the test masses and the satellite. The end-to-end testbed includes all parts of the LTP that can reasonably be tested on earth without free-falling test masses. At its present status it consists mainly of breadboard components. Some of those have already been replaced by engineering models of the LTP experiment. In the next steps, further engineering and flight models will also be inserted in this testbed and tested against well-characterized breadboard components. The presented testbed is an important reference for the unit tests and can also be used for validation of the on-board experiment during the mission.

  5. Flexible end-to-end system design for synthetic aperture radar applications

    NASA Astrophysics Data System (ADS)

    Zaugg, Evan C.; Edwards, Matthew C.; Bradley, Joshua P.

    2012-06-01

    This paper presents ARTEMIS, Inc.'s approach to development of end-to-end synthetic aperture radar systems for multiple applications and platforms. The flexible design of the radar and the image processing tools facilitates their inclusion in a variety of application-specific end-to-end systems. Any given application comes with certain requirements that must be met in order to achieve success. A concept of operation is defined which states how the technology is used to meet the requirements of the application. This drives the design decisions. Key to adapting our system to multiple applications is the flexible SlimSAR radar system, which is programmable on-the-fly to meet the imaging requirements of a wide range of altitudes, swath-widths, and platform velocities. The processing software can be used for real-time imagery production or post-flight processing. The ground station is adaptable, and the radar controls can be run by an operator on the ground, on-board the aircraft, or even automated as part of the aircraft autopilot controls. System integration takes the whole operation into account, seeking to flawlessly work with data links and on-board data storage, aircraft and payload control systems, mission planning, and image processing and exploitation. Examples of applications are presented including using a small unmanned aircraft at low altitude with a line of sight data link, a long-endurance UAV maritime surveillance mission with on-board processing, and a manned ground moving target indicator application with the radar using multiple receive channels.

  6. Laparoscopic Resection and End-to-End Ureteroureterostomy for Midureteral Obstruction in Children.

    PubMed

    Lu, Liangsheng; Bi, Yunli; Wang, Xiang; Ruan, Shuangsui

    2017-02-01

    Midureteral obstruction is an extremely rare condition. This retrospective study assessed the outcomes of laparoscopic ureteroureterostomy (UU) in 13 children with midureteral obstructions. Records of patients with midureteral obstruction who underwent laparoscopic end-to-end UU between July 2011 and August 2015 were reviewed. The medical records comprised patient demographic data, intraoperative details, postoperative outcomes, and related complications. Renal ultrasound, magnetic resonance urography, and radioisotope renography were used for pre- and postoperative assessment. Success was defined as removal of symptomatic obstruction or improvement of hydronephrosis. Thirteen patients from 3.0 months to 12 years old were identified: 6 patients presented with asymptomatic hydronephrosis (according to the renal ultrasonography), 3 with intermittent abdominal pain, and 2 with gross hematuria. The remaining 2 patients presented with urinary extravasation after inguinal hernia repair and the Soave procedure, respectively. All patients underwent laparoscopic end-to-end UU. Procedures were performed using either the laparoscopic transperitoneal approach or laparoscopic retroperitoneal approach. All surgeries were performed successfully except for one case that required open surgery because of difficult anastomosis. There were no intraoperative or immediate postoperative complications. The etiology of the midureteral obstruction included congenital midureteral stricture, ureteral polyp, retrocaval ureter, and iatrogenic injury. The mean surgical time was 148.4 min, and postoperative hospital stay was 4.0-9.0 d. Postoperative outcomes were successful in all 13 patients (100%); median follow-up was 19.1 months. Laparoscopic UU was found to be feasible and safe for the management of midureteral obstruction in children.

  7. Integrating end-to-end threads of control into object-oriented analysis and design

    NASA Technical Reports Server (NTRS)

    Mccandlish, Janet E.; Macdonald, James R.; Graves, Sara J.

    1993-01-01

    Current object-oriented analysis and design methodologies fall short in their use of mechanisms for identifying threads of control for the system being developed. The scenarios which typically describe a system are more global than looking at the individual objects and representing their behavior. Unlike conventional methodologies that use data flow and process-dependency diagrams, object-oriented methodologies do not provide a model for representing these global threads end-to-end. Tracing through threads of control is key to ensuring that a system is complete and timing constraints are addressed. The existence of multiple threads of control in a system necessitates a partitioning of the system into processes. This paper describes the application and representation of end-to-end threads of control to the object-oriented analysis and design process using object-oriented constructs. The issue of representation is viewed as a grouping problem, that is, how to group classes/objects at a higher level of abstraction so that the system may be viewed as a whole with both classes/objects and their associated dynamic behavior. Existing object-oriented development methodology techniques are extended by adding design-level constructs termed logical composite classes and process composite classes. Logical composite classes are design-level classes which group classes/objects both logically and by thread of control information. Process composite classes further refine the logical composite class groupings by using process partitioning criteria to produce optimum concurrent execution results. The goal of these design-level constructs is to ultimately provide the basis for a mechanism that can support the creation of process composite classes in an automated way. Using an automated mechanism makes it easier to partition a system into concurrently executing elements that can be run in parallel on multiple processors.

  8. Early Experience of Robot-Assisted Esophagectomy With Circular End-to-End Stapled Anastomosis.

    PubMed

    Wee, Jon O; Bravo-Iñiguez, Carlos E; Jaklitsch, Michael T

    2016-07-01

    Surgical resection is a critical element in the treatment of esophageal cancer. Esophagectomy is technically challenging and is associated with high morbidity and mortality rates. Efforts to reduce these rates have spurred the adoption of minimally invasive techniques. This study describes a single-institution experience of robot-assisted esophagectomy with circular end-to-end stapled anastomosis. Between December 2013 and April 2015, a series of consecutive patients underwent robot-assisted Ivor Lewis esophagectomy with circular end-to-end anastomosis (RAILE-EEA) at a tertiary care center with curative intent. We retrospectively reviewed their electronic medical records using real-time prospectively collected data. The operative and postoperative outcomes were recorded. Twenty patients underwent RAILE-EEA during the study period. The abdominal mobilization was performed laparoscopically, and the thoracic portion was robotic. The median total operative time was 455 minutes (range, 318-765 minutes), the 90-day operative mortality was 0%, and morbidity was present in 11 of 20 patients (55%). Atrial fibrillation was the most common event and was observed in 3 patients (15%). There were no anastomotic leaks. The median estimated blood loss was 275 mL, and the conversion rate was 0%. Complete (R0) resection was achieved in all cases. The mean number of lymph nodes was 23.2 (± 2.26). The median follow-up time was 330 days (range, 108-600 days), and the overall 1-year survival was 84%. RAILE-EEA in our institution suggests a safe, effective, and reproducible alternative with satisfactory postoperative outcomes for the treatment of esophageal cancer. It provided good local control, adequate lymphadenectomy, low morbidity, and low 90-day operative mortality. Copyright © 2016 The Society of Thoracic Surgeons. Published by Elsevier Inc. All rights reserved.

  9. End-to-end test of spatial accuracy in Gamma Knife treatments for trigeminal neuralgia

    SciTech Connect

    Brezovich, Ivan A. Wu, Xingen; Duan, Jun; Popple, Richard A.; Shen, Sui; Benhabib, Sidi; Huang, Mi; Christian Dobelbower, M.; Fisher III, Winfield S.

    2014-11-01

    Purpose: Spatial accuracy is most crucial when small targets like the trigeminal nerve are treated. Although current quality assurance procedures typically verify that individual apparatus, like the MRI scanner, CT scanner, Gamma Knife, etc., are meeting specifications, the cumulative error of all equipment and procedures combined may exceed safe margins. This study uses an end-to-end approach to assess the overall targeting errors that may have occurred in individual patients previously treated for trigeminal neuralgia. Methods: The trigeminal nerve is simulated by a 3 mm long, 3.175 mm (1/8 in.) diameter MRI-contrast filled cavity embedded within a PMMA plastic capsule. The capsule is positioned within the head frame such that the location of the cavity matches the Gamma Knife coordinates of an arbitrarily chosen, previously treated patient. Gafchromic EBT2 film is placed at the center of the cavity in coronal and sagittal orientations. The films are marked with a pinprick to identify the cavity center. Treatments are planned for radiation delivery with 4 mm collimators according to MRI and CT scans using the clinical localizer boxes and acquisition protocols. Shots are planned so that the 50% isodose surface encompasses the cavity. Following irradiation, the films are scanned and analyzed. Targeting errors are defined as the distance between the pinprick, which represents the intended target, and the centroid of the 50% isodose line, which is the center of the radiation field that was actually delivered. Results: Averaged over ten patient simulations, targeting errors along the x, y, and z coordinates (patient’s left-to-right, posterior-to-anterior, and head-to-foot) were, respectively, −0.060 ± 0.363, −0.350 ± 0.253, and 0.348 ± 0.204 mm when MRI was used for treatment planning. Planning according to CT exhibited generally smaller errors, namely, 0.109 ± 0.167, −0.191 ± 0.144, and 0.211 ± 0.094 mm. The largest errors along individual axes in MRI

  10. MRI simulation: end-to-end testing for prostate radiation therapy using geometric pelvic MRI phantoms

    NASA Astrophysics Data System (ADS)

    Sun, Jidi; Dowling, Jason; Pichler, Peter; Menk, Fred; Rivest-Henault, David; Lambert, Jonathan; Parker, Joel; Arm, Jameen; Best, Leah; Martin, Jarad; Denham, James W.; Greer, Peter B.

    2015-04-01

    To clinically implement MRI simulation or MRI-alone treatment planning requires comprehensive end-to-end testing to ensure an accurate process. The purpose of this study was to design and build a geometric phantom simulating a human male pelvis that is suitable for both CT and MRI scanning and use it to test geometric and dosimetric aspects of MRI simulation including treatment planning and digitally reconstructed radiograph (DRR) generation. A liquid filled pelvic shaped phantom with simulated pelvic organs was scanned in a 3T MRI simulator with dedicated radiotherapy couch-top, laser bridge and pelvic coil mounts. A second phantom with the same external shape but with an internal distortion grid was used to quantify the distortion of the MR image. Both phantoms were also CT scanned as the gold-standard for both geometry and dosimetry. Deformable image registration was used to quantify the MR distortion. Dose comparison was made using a seven-field IMRT plan developed on the CT scan with the fluences copied to the MR image and recalculated using bulk electron densities. Without correction the maximum distortion of the MR compared with the CT scan was 7.5 mm across the pelvis, while this was reduced to 2.6 and 1.7 mm by the vendor’s 2D and 3D correction algorithms, respectively. Within the locations of the internal organs of interest, the distortion was <1.5 and <1 mm with 2D and 3D correction algorithms, respectively. The dose at the prostate isocentre calculated on CT and MRI images differed by 0.01% (1.1 cGy). Positioning shifts were within 1 mm when setup was performed using MRI generated DRRs compared to setup using CT DRRs. The MRI pelvic phantom allows end-to-end testing of the MRI simulation workflow with comparison to the gold-standard CT based process. MRI simulation was found to be geometrically accurate with organ dimensions, dose distributions and DRR based setup within acceptable limits compared to CT.

  11. Fistulotomy with end-to-end primary sphincteroplasty for anal fistula: results from a prospective study.

    PubMed

    Ratto, Carlo; Litta, Francesco; Parello, Angelo; Zaccone, Giuseppe; Donisi, Lorenza; De Simone, Veronica

    2013-02-01

    Fistulotomy plus primary sphincteroplasty for complex anal fistulas is regarded with scepticism, mainly because of the risk of postoperative incontinence. The aim of this study was to evaluate safety and effectiveness of this technique in medium-term follow up and to identify potential predictive factors of success and postoperative continence impairment. This was a prospective observational study conducted at a tertiary care university hospital in Italy. A total of 72 patients with complex anal fistula of cryptoglandular origin underwent fistulotomy and end-to-end primary sphincteroplasty; patients were followed up at 1 week, 1 and 3 months, 1 year, and were invited to participate in a recent follow-up session. Success regarding healing of the fistula was assessed with 3-dimensional endoanal ultrasound and clinical evaluation. Continence status was evaluated using the Cleveland Clinic fecal incontinence score and by patient report of post-defecation soiling. Of the 72 patients, 12 (16.7%) had recurrent fistulas and 29 patients (40.3%) had undergone seton drainage before definitive surgery. At a mean follow-up of 29.4 (SD, 23.7; range, 6-91 months, the success rate of treatment was 95.8% (69 patients). Fistula recurrence was observed in 3 patients at a mean of 17.3 (SD, 10.3; range, 6-26) months of follow-up. Cleveland Clinic fecal incontinence score did not change significantly (p = 0.16). Eight patients (11.6% of those with no baseline incontinence) reported de novo postdefecation soiling. None of the investigated factors was a significant predictor of success. Patients with recurrent fistula after previous fistula surgery had a 5-fold increased probability of having impaired continence (relative risk = 5.00, 95% CI, 1.45-17.27, p = 0.02). The study was limited by potential single-institution bias, lack of anorectal manometry, and lack of quality of life assessment. Fistulotomy with end-to-end primary sphincteroplasty can be considered to be an effective

  12. Changes in end-to-end interactions of tropomyosin affect mouse cardiac muscle dynamics.

    PubMed

    Gaffin, Robert D; Gokulan, Kuppan; Sacchettini, James C; Hewett, Timothy E; Klevitsky, Raisa; Robbins, Jeffrey; Sarin, Vandana; Zawieja, David C; Meininger, Gerald A; Muthuchamy, Mariappan

    2006-08-01

    The ends of striated muscle tropomyosin (TM) are integral for thin filament cooperativity, determining the cooperative unit size and regulating the affinity of TM for actin. We hypothesized that altering the alpha-TM carboxy terminal overlap end to the beta-TM counterpart would affect the amino-terminal association, which would alter the end-to-end interactions of TM molecules in the thin filament regulatory strand and affect the mechanisms of cardiac muscle contraction. To test this hypothesis, we generated transgenic (TG) mouse lines that express a mutant form of alpha-TM in which the first 275 residues are from alpha-TM and the last nine amino acids are from beta-TM (alpha-TM9aaDeltabeta). Molecular analyses show that endogenous alpha-TM mRNA and protein are nearly completely replaced with alpha-TM9aaDeltabeta. Working heart preparations data show that the rates of contraction and relaxation are reduced in alpha-TM9aaDeltabeta hearts. Left ventricular pressure and time to peak pressure are also reduced (-12% and -13%, respectively). The ratio of maximum to minimum first derivatives of change in left ventricular systolic pressure with respect to time (ratio of +dP/dt to -dP/dt, respectively) is increased, but tau is not changed significantly. Force-intracellular calcium concentration ([Ca2+]i) measurements from intact papillary fibers demonstrate that alpha-TM9aaDeltabeta TG fibers produce less force per given [Ca2+]i compared with nontransgenic fibers. Taken together, the data demonstrate that the rate of contraction is primarily affected in TM TG hearts. Protein docking studies show that in the mutant molecule, the overall carbon backbone is perturbed about 1.5 A, indicating that end-to-end interactions are altered. These results demonstrate that the localized flexibility present in the coiled-coil structures of TM isoforms is different, and that plays an important role in interacting with neighboring thin filament regulatory proteins and with differentially

  13. Availability and End-to-end Reliability in Low Duty Cycle Multihop Wireless Sensor Networks.

    PubMed

    Suhonen, Jukka; Hämäläinen, Timo D; Hännikäinen, Marko

    2009-01-01

    A wireless sensor network (WSN) is an ad-hoc technology that may even consist of thousands of nodes, which necessitates autonomic, self-organizing and multihop operations. A typical WSN node is battery powered, which makes the network lifetime the primary concern. The highest energy efficiency is achieved with low duty cycle operation, however, this alone is not enough. WSNs are deployed for different uses, each requiring acceptable Quality of Service (QoS). Due to the unique characteristics of WSNs, such as dynamic wireless multihop routing and resource constraints, the legacy QoS metrics are not feasible as such. We give a new definition to measure and implement QoS in low duty cycle WSNs, namely availability and reliability. Then, we analyze the effect of duty cycling for reaching the availability and reliability. The results are obtained by simulations with ZigBee and proprietary TUTWSN protocols. Based on the results, we also propose a data forwarding algorithm suitable for resource constrained WSNs that guarantees end-to-end reliability while adding a small overhead that is relative to the packet error rate (PER). The forwarding algorithm guarantees reliability up to 30% PER.

  14. SPOKES: An end-to-end simulation facility for spectroscopic cosmological surveys

    NASA Astrophysics Data System (ADS)

    Nord, B.; Amara, A.; Réfrégier, A.; Gamper, La.; Gamper, Lu.; Hambrecht, B.; Chang, C.; Forero-Romero, J. E.; Serrano, S.; Cunha, C.; Coles, O.; Nicola, A.; Busha, M.; Bauer, A.; Saunders, W.; Jouvel, S.; Kirk, D.; Wechsler, R.

    2016-04-01

    The nature of dark matter, dark energy and large-scale gravity pose some of the most pressing questions in cosmology today. These fundamental questions require highly precise measurements, and a number of wide-field spectroscopic survey instruments are being designed to meet this requirement. A key component in these experiments is the development of a simulation tool to forecast science performance, define requirement flow-downs, optimize implementation, demonstrate feasibility, and prepare for exploitation. We present SPOKES (SPectrOscopic KEn Simulation), an end-to-end simulation facility for spectroscopic cosmological surveys designed to address this challenge. SPOKES is based on an integrated infrastructure, modular function organization, coherent data handling and fast data access. These key features allow reproducibility of pipeline runs, enable ease of use and provide flexibility to update functions within the pipeline. The cyclic nature of the pipeline offers the possibility to make the science output an efficient measure for design optimization and feasibility testing. We present the architecture, first science, and computational performance results of the simulation pipeline. The framework is general, but for the benchmark tests, we use the Dark Energy Spectrometer (DESpec), one of the early concepts for the upcoming project, the Dark Energy Spectroscopic Instrument (DESI). We discuss how the SPOKES framework enables a rigorous process to optimize and exploit spectroscopic survey experiments in order to derive high-precision cosmological measurements optimally.

  15. End-to-end differential contactless conductivity sensor for microchip capillary electrophoresis.

    PubMed

    Fercher, Georg; Haller, Anna; Smetana, Walter; Vellekoop, Michael J

    2010-04-15

    In this contribution, a novel measurement approach for miniaturized capillary electrophoresis (CE) devices is presented: End-to-end differential capacitively coupled contactless conductivity measurement. This measurement technique is applied to a miniaturized CE device fabricated in low-temperature cofired ceramics (LTCC) multilayer technology. The working principle is based on the placement of two distinct detector areas near both ends of the fluid inlet and outlet of the separation channel. Both output signals are subtracted from each other, and the resulting differential signal is amplified and measured. This measurement approach has several advantages over established, single-end detectors: The high baseline level resulting from parasitic stray capacitance and buffer conductivity is reduced, leading to better signal-to-noise ratio and hence higher measurement sensitivity. Furthermore, temperature and, thus, baseline drift effects are diminished owing to the differentiating nature of the system. By comparing the peak widths measured with both detectors, valuable information about zone dispersion effects arising during the separation is obtained. Additionally, the novel measurement scheme allows the determination of dispersion effects that occur at the time of sample injection. Optical means of dispersion evaluation are ineffective because of the opaque LTCC substrate. Electrophoretic separation experiments of inorganic ions show sensitivity enhancements by about a factor of 30-60 compared to the single-end measurement scheme.

  16. End-to-end performance modeling of passive remote sensing systems

    SciTech Connect

    Smith, B.W.; Borel, C.C.; Clodius, W.B.; Theiler, J.; Laubscher, B.; Weber, P.G.

    1996-07-01

    The ultimate goal of end-to-end system modeling is to simulate all known physical effects which determine the content of the data, before flying an instrument system. In remote sensing, one begins with a scene, viewed either statistically or dynamically, computes the radiance in each spectral band, renders the scene, transfers it through representative atmospheres to create the radiance field at an aperture, and integrates over sensor pixels. We have simulated a comprehensive sequence of realistic instrument hardware elements and the transfer of simulated data to an analysis system. This analysis package is the same as that intended for use of data collections from the real system. By comparing the analyzed image to the original scene, the net effect of nonideal system components can be understood. Iteration yields the optimum values of system parameters to achieve performance targets. We have used simulation to develop and test improved multispectral algorithms for (1) the robust retrieval of water surface temperature, water vapor column, and other quantities; (2) the preservation of radiometric accuracy during atmospheric correction and pixel registration on the ground; and (3) exploitation of on-board multispectral measurements to assess the atmosphere between ground and aperture.

  17. End-to-End Optimization of High-Throughput DNA Sequencing.

    PubMed

    O'Reilly, Eliza; Baccelli, Francois; De Veciana, Gustavo; Vikalo, Haris

    2016-10-01

    At the core of Illumina's high-throughput DNA sequencing platforms lies a biophysical surface process that results in a random geometry of clusters of homogeneous short DNA fragments typically hundreds of base pairs long-bridge amplification. The statistical properties of this random process and the lengths of the fragments are critical as they affect the information that can be subsequently extracted, that is, density of successfully inferred DNA fragment reads. The ensembles of overlapping DNA fragment reads are then used to computationally reconstruct the much longer target genome sequence. The success of the reconstruction in turn depends on having a sufficiently large ensemble of DNA fragments that are sufficiently long. In this article using stochastic geometry, we model and optimize the end-to-end flow cell synthesis and target genome sequencing process, linking and partially controlling the statistics of the physical processes to the success of the final computational step. Based on a rough calibration of our model, we provide, for the first time, a mathematical framework capturing the salient features of the sequencing platform that serves as a basis for optimizing cost, performance, and/or sensitivity analysis to various parameters.

  18. Semantic Complex Event Processing over End-to-End Data Flows

    SciTech Connect

    Zhou, Qunzhi; Simmhan, Yogesh; Prasanna, Viktor K.

    2012-04-01

    Emerging Complex Event Processing (CEP) applications in cyber physical systems like SmartPower Grids present novel challenges for end-to-end analysis over events, flowing from heterogeneous information sources to persistent knowledge repositories. CEP for these applications must support two distinctive features - easy specification patterns over diverse information streams, and integrated pattern detection over realtime and historical events. Existing work on CEP has been limited to relational query patterns, and engines that match events arriving after the query has been registered. We propose SCEPter, a semantic complex event processing framework which uniformly processes queries over continuous and archived events. SCEPteris built around an existing CEP engine with innovative support for semantic event pattern specification and allows their seamless detection over past, present and future events. Specifically, we describe a unified semantic query model that can operate over data flowing through event streams to event repositories. Compile-time and runtime semantic patterns are distinguished and addressed separately for efficiency. Query rewriting is examined and analyzed in the context of temporal boundaries that exist between event streams and their repository to avoid duplicate or missing results. The design and prototype implementation of SCEPterare analyzed using latency and throughput metrics for scenarios from the Smart Grid domain.

  19. An end-to-end assessment of range uncertainty in proton therapy using animal tissues

    NASA Astrophysics Data System (ADS)

    Zheng, Yuanshui; Kang, Yixiu; Zeidan, Omar; Schreuder, Niek

    2016-11-01

    Accurate assessment of range uncertainty is critical in proton therapy. However, there is a lack of data and consensus on how to evaluate the appropriate amount of uncertainty. The purpose of this study is to quantify the range uncertainty in various treatment conditions in proton therapy, using transmission measurements through various animal tissues. Animal tissues, including a pig head, beef steak, and lamb leg, were used in this study. For each tissue, an end-to-end test closely imitating patient treatments was performed. This included CT scan simulation, treatment planning, image-guided alignment, and beam delivery. Radio-chromic films were placed at various depths in the distal dose falloff region to measure depth dose. Comparisons between measured and calculated doses were used to evaluate range differences. The dose difference at the distal falloff between measurement and calculation depends on tissue type and treatment conditions. The estimated range difference was up to 5, 6 and 4 mm for the pig head, beef steak, and lamb leg irradiation, respectively. Our study shows that the TPS was able to calculate proton range within about 1.5% plus 1.5 mm. Accurate assessment of range uncertainty in treatment planning would allow better optimization of proton beam treatment, thus fully achieving proton beams’ superior dose advantage over conventional photon-based radiation therapy.

  20. An end-to-end analysis of drought from smallholder farms in southwest Jamaica

    NASA Astrophysics Data System (ADS)

    Curtis, W. R. S., III; Gamble, D. W.; Popke, J.

    2015-12-01

    Drought can be defined in many ways: meteorological, hydrological, agricultural, and socio-economic. Another way to approach drought is from a "perception" perspective, where individuals whose livelihood is highly dependent on precipitation take adaptive actions. In this study we use two-years of data collected from twelve smallholder farms in southern St. Elizabeth, Jamaica to undertake an end-to-end analysis of drought. At each farm, 6-hour temperature and soil moisture, and tipping-bucket rainfall were recorded from June 2013 to June 2015, and twice-monthly farmers indicated whether they were experiencing drought and if they irrigated (hand-watering, drip irrigation, or pipe and sprinkler). In many cases half of the farmers considered themselves in a drought, while the others not, even though the largest separation among farms was about 20 km. This study will use analysis of variance to test the following hypotheses: Drought perception is related to a) absolute amounts of precipitation at the time, b) other environmental cues at the time (soil moisture, temperature), or c) relative amounts of precipitation as compared to the same time last year. Irrigation actions and water use following the perception of drought will also be examined.

  1. Availability and End-to-end Reliability in Low Duty Cycle Multihop Wireless Sensor Networks

    PubMed Central

    Suhonen, Jukka; Hämäläinen, Timo D.; Hännikäinen, Marko

    2009-01-01

    A wireless sensor network (WSN) is an ad-hoc technology that may even consist of thousands of nodes, which necessitates autonomic, self-organizing and multihop operations. A typical WSN node is battery powered, which makes the network lifetime the primary concern. The highest energy efficiency is achieved with low duty cycle operation, however, this alone is not enough. WSNs are deployed for different uses, each requiring acceptable Quality of Service (QoS). Due to the unique characteristics of WSNs, such as dynamic wireless multihop routing and resource constraints, the legacy QoS metrics are not feasible as such. We give a new definition to measure and implement QoS in low duty cycle WSNs, namely availability and reliability. Then, we analyze the effect of duty cycling for reaching the availability and reliability. The results are obtained by simulations with ZigBee and proprietary TUTWSN protocols. Based on the results, we also propose a data forwarding algorithm suitable for resource constrained WSNs that guarantees end-to-end reliability while adding a small overhead that is relative to the packet error rate (PER). The forwarding algorithm guarantees reliability up to 30% PER. PMID:22574002

  2. An end-to-end assessment of range uncertainty in proton therapy using animal tissues.

    PubMed

    Zheng, Yuanshui; Kang, Yixiu; Zeidan, Omar; Schreuder, Niek

    2016-11-21

    Accurate assessment of range uncertainty is critical in proton therapy. However, there is a lack of data and consensus on how to evaluate the appropriate amount of uncertainty. The purpose of this study is to quantify the range uncertainty in various treatment conditions in proton therapy, using transmission measurements through various animal tissues. Animal tissues, including a pig head, beef steak, and lamb leg, were used in this study. For each tissue, an end-to-end test closely imitating patient treatments was performed. This included CT scan simulation, treatment planning, image-guided alignment, and beam delivery. Radio-chromic films were placed at various depths in the distal dose falloff region to measure depth dose. Comparisons between measured and calculated doses were used to evaluate range differences. The dose difference at the distal falloff between measurement and calculation depends on tissue type and treatment conditions. The estimated range difference was up to 5, 6 and 4 mm for the pig head, beef steak, and lamb leg irradiation, respectively. Our study shows that the TPS was able to calculate proton range within about 1.5% plus 1.5 mm. Accurate assessment of range uncertainty in treatment planning would allow better optimization of proton beam treatment, thus fully achieving proton beams' superior dose advantage over conventional photon-based radiation therapy.

  3. Vascular Coupling System for End-to-End Anastomosis: An In Vivo Pilot Case Report.

    PubMed

    Li, Huizhong; Gale, Bruce; Shea, Jill; Sant, Himanshu; Terry, Christi M; Agarwal, Jay

    2017-03-01

    This paper presents the latest in vivo findings of a novel vascular coupling system. Vascular anastomosis is a common procedure in reconstructive surgeries and traditional hand suturing is very time consuming. The vascular coupling system described herein was designed to be used on arteries for a rapid and error-free anastomosis. The system consists of an engaging ring made from high density polyethylene using computer numerical control machining and a back ring made from polymethylmethacrylate using laser cutting. The vascular coupling system and its corresponding installation tools were tested in a pilot animal study to evaluate their efficacy in completing arterial anastomosis. A segment of expanded polytetrafluoroethylene (ePTFE) tubing was interposed into a transected carotid artery by anastomosis using two couplers in a pig. Two end-to-end anastomoses were accomplished. Ultrasound images were obtained to evaluate the blood flow at the anastomotic site immediately after the surgery. MRI was also performed 2 weeks after the surgery to evaluate vessel and ePTFE graft patency. This anastomotic system demonstrated high efficacy and easy usability, which should facilitate vascular anastomosis procedures in trauma and reconstructive surgeries.

  4. The X-IFU end-to-end simulations performed for the TES array optimization exercise

    NASA Astrophysics Data System (ADS)

    Peille, Philippe; Wilms, J.; Brand, T.; Cobo, B.; Ceballos, M. T.; Dauser, T.; Smith, S. J.; Barret, D.; den Herder, J. W.; Piro, L.; Barcons, X.; Pointecouteau, E.; Bandler, S.; den Hartog, R.; de Plaa, J.

    2015-09-01

    The focal plane assembly of the Athena X-ray Integral Field Unit (X-IFU) includes as the baseline an array of ~4000 single size calorimeters based on Transition Edge Sensors (TES). Other sensor array configurations could however be considered, combining TES of different properties (e.g. size). In attempting to improve the X-IFU performance in terms of field of view, count rate performance, and even spectral resolution, two alternative TES array configurations to the baseline have been simulated, each combining a small and a large pixel array. With the X-IFU end-to-end simulator, a sub-sample of the Athena core science goals, selected by the X-IFU science team as potentially driving the optimal TES array configuration, has been simulated for the results to be scientifically assessed and compared. In this contribution, we will describe the simulation set-up for the various array configurations, and highlight some of the results of the test cases simulated.

  5. End-To-End Simulation of Launch Vehicle Trajectories Including Stage Separation Dynamics

    NASA Technical Reports Server (NTRS)

    Albertson, Cindy W.; Tartabini, Paul V.; Pamadi, Bandu N.

    2012-01-01

    The development of methodologies, techniques, and tools for analysis and simulation of stage separation dynamics is critically needed for successful design and operation of multistage reusable launch vehicles. As a part of this activity, the Constraint Force Equation (CFE) methodology was developed and implemented in the Program to Optimize Simulated Trajectories II (POST2). The objective of this paper is to demonstrate the capability of POST2/CFE to simulate a complete end-to-end mission. The vehicle configuration selected was the Two-Stage-To-Orbit (TSTO) Langley Glide Back Booster (LGBB) bimese configuration, an in-house concept consisting of a reusable booster and an orbiter having identical outer mold lines. The proximity and isolated aerodynamic databases used for the simulation were assembled using wind-tunnel test data for this vehicle. POST2/CFE simulation results are presented for the entire mission, from lift-off, through stage separation, orbiter ascent to orbit, and booster glide back to the launch site. Additionally, POST2/CFE stage separation simulation results are compared with results from industry standard commercial software used for solving dynamics problems involving multiple bodies connected by joints.

  6. Forming End-to-End Oligomers of Gold Nanorods Using Porphyrins and Phthalocyanines.

    PubMed

    Stewart, Alexander F; Gagnon, Brandon P; Walker, Gilbert C

    2015-06-23

    The illumination of aggregated metal nanospecies can create strong local electric fields to brighten Raman scattering. This study describes a procedure to self-assemble gold nanorods (NRs) through the use of porphyrin and phthalocyanine agents to create reproducibly stable and robust NR aggregates in the form of end-to-end oligomers. Narrow inter-rod gaps result, creating electric field "hot spots" between the NRs. The organic linker molecules themselves are potential Raman-based optical labels, and the result is significant numbers of Raman-active species located in the hot spots. NR polymerization was quenched by phospholipid encapsulation, which allows for control of the polydispersity of the aggregate solution, to optimize the surface-enhanced Raman scattering (SERS) enhancement and permitted the aqueous solubility of the aggregates. The increased presence of Raman-active species in the hot spots and the optimizing of solution polydispersity resulted in the observation of scattering enhancements by encapsulated porphyrins/phthalocyanines of up to 3500-fold over molecular chromophores lacking the NR oligomer host.

  7. An end-to-end architecture for distributing weather alerts to wireless handsets

    NASA Astrophysics Data System (ADS)

    Jones, Karen L.; Nguyen, Hung

    2005-06-01

    This paper describes the current National Weather Service's (NWS) system for providing weather alerts in the U.S. and will review how the existing end-to-end architecture is being leveraged to provide non-weather alerts, also known as "all-hazard alerts", to the general public. The paper then describes how a legacy system that transmits weather and all-hazard alerts can be extended via commercial wireless networks and protocols to reach 154 million Americans who carry cell phones. This approach uses commercial SATCOM and existing wireless carriers and services such as Short Messaging Service (SMS) for text and emerging Multimedia Messaging Service (MMS) protocol, which would allow for photos, maps, audio and video alerts to be sent to end users. This wireless broadcast alert delivery architecture is designed to be open and to embrace the National Weather Service's mandate to become an "" warning system for the general public. Examples of other public and private sector applications that require timely and intelligent push mechanisms using this alert dissemination approach are also given.

  8. SPOKES: An end-to-end simulation facility for spectroscopic cosmological surveys

    DOE PAGES

    Nord, B.; Amara, A.; Refregier, A.; ...

    2016-03-03

    The nature of dark matter, dark energy and large-scale gravity pose some of the most pressing questions in cosmology today. These fundamental questions require highly precise measurements, and a number of wide-field spectroscopic survey instruments are being designed to meet this requirement. A key component in these experiments is the development of a simulation tool to forecast science performance, define requirement flow-downs, optimize implementation, demonstrate feasibility, and prepare for exploitation. We present SPOKES (SPectrOscopic KEn Simulation), an end-to-end simulation facility for spectroscopic cosmological surveys designed to address this challenge. SPOKES is based on an integrated infrastructure, modular function organization, coherentmore » data handling and fast data access. These key features allow reproducibility of pipeline runs, enable ease of use and provide flexibility to update functions within the pipeline. The cyclic nature of the pipeline offers the possibility to make the science output an efficient measure for design optimization and feasibility testing. We present the architecture, first science, and computational performance results of the simulation pipeline. The framework is general, but for the benchmark tests, we use the Dark Energy Spectrometer (DESpec), one of the early concepts for the upcoming project, the Dark Energy Spectroscopic Instrument (DESI). As a result, we discuss how the SPOKES framework enables a rigorous process to optimize and exploit spectroscopic survey experiments in order to derive high-precision cosmological measurements optimally.« less

  9. New Development in WARP: Progress Toward End-to-End Simulation

    SciTech Connect

    Grote, D.P.; Friedman, A.; Haber, I.; Fawley, W.; Luc Vay, J.

    1997-11-07

    The development of a high current, heavy-ion beam driver for inertial confinement fusion requires a detailed understanding of the behavior of the beam, including effects of the strong self-fields. The necessity of including the self-fields of the beam makes particle-in-cell (PIC) simulation techniques ideal, and for this reason, the multi-dimensional PIC/accelerator code WARP has been developed. WARP has been used extensively to study the creation and propagation of ion beams both in experiments and for the understanding of basic beam physics. An overview of the structure of the code will be presented along with a discussion of features that make the code an effective tool in the understanding of space-charge dominated beam behavior. Much development has been done on WARP increasing its flexibility and generality. Major additions include a generalized field description, an efficient steady-state modelling technique, a transverse slice model with a bending algorithm, further improvement of the parallel processing version, and capabilities for linking to chamber transport codes. With these additions, the capability of modeling a large scale accelerator from end-to-end comes closer to reality.

  10. End-to-end simulation of bunch merging for a muon collider

    SciTech Connect

    Bao, Yu; Stratakis, Diktys; Hanson, Gail G.; Palmer, Robert B.

    2015-05-03

    Muon accelerator beams are commonly produced indirectly through pion decay by interaction of a charged particle beam with a target. Efficient muon capture requires the muons to be first phase-rotated by rf cavities into a train of 21 bunches with much reduced energy spread. Since luminosity is proportional to the square of the number of muons per bunch, it is crucial for a Muon Collider to use relatively few bunches with many muons per bunch. In this paper we will describe a bunch merging scheme that should achieve this goal. We present for the first time a complete end-to-end simulation of a 6D bunch merger for a Muon Collider. The 21 bunches arising from the phase-rotator, after some initial cooling, are merged in longitudinal phase space into seven bunches, which then go through seven paths with different lengths and reach the final collecting "funnel" at the same time. The final single bunch has a transverse and a longitudinal emittance that matches well with the subsequent 6D rectilinear cooling scheme.

  11. SPOKES: An end-to-end simulation facility for spectroscopic cosmological surveys

    SciTech Connect

    Nord, B.; Amara, A.; Refregier, A.; Gamper, La.; Gamper, Lu.; Hambrecht, B.; Chang, C.; Forero-Romero, J. E.; Serrano, S.; Cunha, C.; Coles, O.; Nicola, A.; Busha, M.; Bauer, A.; Saunders, W.; Jouvel, S.; Kirk, D.; Wechsler, R.

    2016-03-03

    The nature of dark matter, dark energy and large-scale gravity pose some of the most pressing questions in cosmology today. These fundamental questions require highly precise measurements, and a number of wide-field spectroscopic survey instruments are being designed to meet this requirement. A key component in these experiments is the development of a simulation tool to forecast science performance, define requirement flow-downs, optimize implementation, demonstrate feasibility, and prepare for exploitation. We present SPOKES (SPectrOscopic KEn Simulation), an end-to-end simulation facility for spectroscopic cosmological surveys designed to address this challenge. SPOKES is based on an integrated infrastructure, modular function organization, coherent data handling and fast data access. These key features allow reproducibility of pipeline runs, enable ease of use and provide flexibility to update functions within the pipeline. The cyclic nature of the pipeline offers the possibility to make the science output an efficient measure for design optimization and feasibility testing. We present the architecture, first science, and computational performance results of the simulation pipeline. The framework is general, but for the benchmark tests, we use the Dark Energy Spectrometer (DESpec), one of the early concepts for the upcoming project, the Dark Energy Spectroscopic Instrument (DESI). As a result, we discuss how the SPOKES framework enables a rigorous process to optimize and exploit spectroscopic survey experiments in order to derive high-precision cosmological measurements optimally.

  12. End-To-End performance test of the LINC-NIRVANA Wavefront-Sensor system.

    NASA Astrophysics Data System (ADS)

    Berwein, Juergen; Bertram, Thomas; Conrad, Al; Briegel, Florian; Kittmann, Frank; Zhang, Xiangyu; Mohr, Lars

    2011-09-01

    LINC-NIRVANA is an imaging Fizeau interferometer, for use in near infrared wavelengths, being built for the Large Binocular Telescope. Multi-conjugate adaptive optics (MCAO) increases the sky coverage and the field of view over which diffraction limited images can be obtained. For its MCAO implementation, Linc-Nirvana utilizes four total wavefront sensors; each of the two beams is corrected by both a ground-layer wavefront sensor (GWS) and a high-layer wavefront sensor (HWS). The GWS controls the adaptive secondary deformable mirror (DM), which is based on an DSP slope computing unit. Whereas the HWS controls an internal DM via computations provided by an off-the-shelf multi-core Linux system. Using wavefront sensor data collected from a prior lab experiment, we have shown via simulation that the Linux based system is sufficient to operate at 1kHz, with jitter well below the needs of the final system. Based on that setup we tested the end-to-end performance and latency through all parts of the system which includes the camera, the wavefront controller, and the deformable mirror. We will present our loop control structure and the results of those performance tests.

  13. Direct end-to-end repair of flexor pollicis longus tendon lacerations.

    PubMed

    Nunley, J A; Levin, L S; Devito, D; Goldner, R D; Urbaniak, J R

    1992-01-01

    Between 1976 and 1986, 38 consecutive acute isolated flexor pollicis longus lacerations were repaired. This study excluded all replanted or mutilated digits and all lacerations with associated fracture. Average follow-up was 26 months. Tendon rehabilitation was standardized. Range of motion and pinch strength were measured postoperatively. Seventy-four percent (28/38) of the flexor pollicis longus injuries occurred in zone II. Neurovascular injury occurred in 82% of the lacerations, and this correlated with the zone of tendon injury. In 21% of the patients (8/38) both digital nerves and arteries were transected. Postoperative thumb interphalangeal motion averaged 35 degrees and key pinch strength was 81% that of the uninjured thumb. One rupture occurred in a child. Laceration of the flexor pollicis longus is likely to involve damage to neurovascular structures, and repair may be necessary. Direct end-to-end repairs within the pulley system do at least as well as delayed tendon reconstruction and do not require additional procedures.

  14. The optical performance of the PILOT instrument from ground end-to-end tests

    NASA Astrophysics Data System (ADS)

    Misawa, R.; Bernard, J.-Ph.; Longval, Y.; Ristorcelli, I.; Ade, P.; Alina, D.; André, Y.; Aumont, J.; Bautista, L.; de Bernardis, P.; Boulade, O.; Bousqet, F.; Bouzit, M.; Buttice, V.; Caillat, A.; Chaigneau, M.; Charra, M.; Crane, B.; Douchin, F.; Doumayrou, E.; Dubois, J. P.; Engel, C.; Griffin, M.; Foenard, G.; Grabarnik, S.; Hargrave, P.; Hughes, A.; Laureijs, R.; Leriche, B.; Maestre, S.; Maffei, B.; Marty, C.; Marty, W.; Masi, S.; Montel, J.; Montier, L.; Mot, B.; Narbonne, J.; Pajot, F.; Pérot, E.; Pimentao, J.; Pisano, G.; Ponthieu, N.; Rodriguez, L.; Roudil, G.; Salatino, M.; Savini, G.; Simonella, O.; Saccoccio, M.; Tauber, J.; Tucker, C.

    2017-06-01

    The Polarized Instrument for Long-wavelength Observation of the Tenuous interstellar medium ( PILOT) is a balloon-borne astronomy experiment designed to study the linear polarization of thermal dust emission in two photometric bands centred at wavelengths 240 μm (1.2 THz) and 550 μm (545 GHz), with an angular resolution of a few arcminutes. Several end-to-end tests of the instrument were performed on the ground between 2012 and 2014, in order to prepare for the first scientific flight of the experiment that took place in September 2015 from Timmins, Ontario, Canada. This paper presents the results of those tests, focussing on an evaluation of the instrument's optical performance. We quantify image quality across the extent of the focal plane, and describe the tests that we conducted to determine the focal plane geometry, the optimal focus position, and sources of internal straylight. We present estimates of the detector response, obtained using an internal calibration source, and estimates of the background intensity and background polarization.

  15. Development of an End-to-End Model for Free-Space Optical Communications

    NASA Astrophysics Data System (ADS)

    Hemmati, H.

    2005-05-01

    Through funding by NASA's Exploration Systems Research and Technology (ESR&T) Program and the Advanced Space Technology Program (ASTP), a team, including JPL, Boeing, NASA-Glenn, and the Georgia Institute of Technology, will develop an end-to-end modeling tool for rapid architecture trade-offs of high-data-rate laser communications from lunar, martian, and outer planetary ranges. An objective of the modeling tool is to reduce the inefficient reliance on modeling of discrete subsystems or sequential development of multiple expensive and time-consuming hardware units, thereby saving significant cost and time. This dynamic, time-domain modeling tool will accept measured component and subsystem data inputs and generate "difficult to measure" characteristics required for the performance evaluation of different designs and architectural choices. The planned modeling tool will incorporate actual subsystem performance data to reduce the develop-build-evaluate-refine production cycle. The list of high-level objectives of the program includes (1) development of a bidirectional global link analysis backbone software encompassing all optical communication subsystem parameters; (2) development of a bidirectional global link simulation model encompassing all optical communication parameters; (3) interoperability of the link analysis tool with all relevant detailed subsystem design models; and (4) a validated model that is validated against known experimental data at the subsystem and system levels.

  16. Towards a cross-platform software framework to support end-to-end hydrometeorological sensor network deployment

    NASA Astrophysics Data System (ADS)

    Celicourt, P.; Sam, R.; Piasecki, M.

    2016-12-01

    Global phenomena such as climate change and large scale environmental degradation require the collection of accurate environmental data at detailed spatial and temporal scales from which knowledge and actionable insights can be derived using data science methods. Despite significant advances in sensor network technologies, sensors and sensor network deployment remains a labor-intensive, time consuming, cumbersome and expensive task. These factors demonstrate why environmental data collection remains a challenge especially in developing countries where technical infrastructure, expertise and pecuniary resources are scarce. In addition, they also demonstrate the reason why dense and long-term environmental data collection has been historically quite difficult. Moreover, hydrometeorological data collection efforts usually overlook the (critically important) inclusion of a standards-based system for storing, managing, organizing, indexing, documenting and sharing sensor data. We are developing a cross-platform software framework using the Python programming language that will allow us to develop a low cost end-to-end (from sensor to publication) system for hydrometeorological conditions monitoring. The software framework contains provision for sensor, sensor platforms, calibration and network protocols description, sensor programming, data storage, data publication and visualization and more importantly data retrieval in a desired unit system. It is being tested on the Raspberry Pi microcomputer as end node and a laptop PC as the base station in a wireless setting.

  17. Status report of the end-to-end ASKAP software system: towards early science operations

    NASA Astrophysics Data System (ADS)

    Guzman, Juan Carlos; Chapman, Jessica; Marquarding, Malte; Whiting, Matthew

    2016-08-01

    300 MHz bandwidth for Array Release 1; followed by the deployment of the real-time data processing components. In addition to the Central Processor, the first production release of the CSIRO ASKAP Science Data Archive (CASDA) has also been deployed in one of the Pawsey Supercomputing Centre facilities and it is integrated to the end-to-end ASKAP data flow system. This paper describes the current status of the "end-to-end" data flow software system from preparing observations to data acquisition, processing and archiving; and the challenges of integrating an HPC facility as a key part of the instrument. It also shares some lessons learned since the start of integration activities and the challenges ahead in preparation for the start of the Early Science program.

  18. SPoRT - An End-to-End R2O Activity

    NASA Technical Reports Server (NTRS)

    Jedlovec, Gary J.

    2009-01-01

    Established in 2002 to demonstrate the weather and forecasting application of real-time EOS measurements, the Short-term Prediction Research and Transition (SPoRT) program has grown to be an end-to-end research to operations activity focused on the use of advanced NASA modeling and data assimilation approaches, nowcasting techniques, and unique high-resolution multispectral observational data applications from EOS satellites to improve short-term weather forecasts on a regional and local scale. SPoRT currently partners with several universities and other government agencies for access to real-time data and products, and works collaboratively with them and operational end users at 13 WFOs to develop and test the new products and capabilities in a "test-bed" mode. The test-bed simulates key aspects of the operational environment without putting constraints on the forecaster workload. Products and capabilities which show utility in the test-bed environment are then transitioned experimentally into the operational environment for further evaluation and assessment. SPoRT focuses on a suite of data and products from MODIS, AMSR-E, and AIRS on the NASA Terra and Aqua satellites, and total lightning measurements from ground-based networks. Some of the observations are assimilated into or used with various versions of the WRF model to provide supplemental forecast guidance to operational end users. SPoRT is enhancing partnerships with NOAA / NESDIS for new product development and data access to exploit the remote sensing capabilities of instruments on the NPOESS satellites to address short term weather forecasting problems. The VIIRS and CrIS instruments on the NPP and follow-on NPOESS satellites provide similar observing capabilities to the MODIS and AIRS instruments on Terra and Aqua. SPoRT will be transitioning existing and new capabilities into the AWIIPS II environment to continue the continuity of its activities.

  19. SPoRT - An End-to-End R2O Activity

    NASA Technical Reports Server (NTRS)

    Jedlovec, Gary J.

    2009-01-01

    Established in 2002 to demonstrate the weather and forecasting application of real-time EOS measurements, the Short-term Prediction Research and Transition (SPoRT) program has grown to be an end-to-end research to operations activity focused on the use of advanced NASA modeling and data assimilation approaches, nowcasting techniques, and unique high-resolution multispectral observational data applications from EOS satellites to improve short-term weather forecasts on a regional and local scale. SPoRT currently partners with several universities and other government agencies for access to real-time data and products, and works collaboratively with them and operational end users at 13 WFOs to develop and test the new products and capabilities in a "test-bed" mode. The test-bed simulates key aspects of the operational environment without putting constraints on the forecaster workload. Products and capabilities which show utility in the test-bed environment are then transitioned experimentally into the operational environment for further evaluation and assessment. SPoRT focuses on a suite of data and products from MODIS, AMSR-E, and AIRS on the NASA Terra and Aqua satellites, and total lightning measurements from ground-based networks. Some of the observations are assimilated into or used with various versions of the WRF model to provide supplemental forecast guidance to operational end users. SPoRT is enhancing partnerships with NOAA / NESDIS for new product development and data access to exploit the remote sensing capabilities of instruments on the NPOESS satellites to address short term weather forecasting problems. The VIIRS and CrIS instruments on the NPP and follow-on NPOESS satellites provide similar observing capabilities to the MODIS and AIRS instruments on Terra and Aqua. SPoRT will be transitioning existing and new capabilities into the AWIIPS II environment to continue the continuity of its activities.

  20. End-to-End Models for Effects of System Noise on LIMS Analysis of Igneous Rocks

    SciTech Connect

    Clegg, Samuel M; Bender, Steven; Wiens, R. C.; Carmosino, Marco L; Speicher, Elly A; Dyar, M. D.

    2010-12-23

    The ChemCam instrument on the Mars Science Laboratory will be the first extraterrestial deployment of laser-induced breakdown spectroscopy (UBS) for remote geochemical analysis. LIBS instruments are also being proposed for future NASA missions. In quantitative LIBS applications using multivariate analysis techniques, it is essential to understand the effects of key instrument parameters and their variability on the elemental predictions. Baseline experiments were run on a laboratory instrument in conditions reproducing ChemCam performance on Mars. These experiments employed Nd:YAG laser producing 17 mJ/pulse on target and an with a 200 {micro}m FWHM spot size on the surface of a sample. The emission is collected by a telescope, imaged on a fiber optic and then interfaced to a demultiplexer capable of >40% transmission into each spectrometer. We report here on an integrated end-to-end system performance model that simulates the effects of output signal degradation that might result from the input signal chain and the impact on multivariate model predictions. There are two approaches to modifying signal to noise (SNR): degrade the signal and/or increase the noise. Ishibashi used a much smaller data set to show that the addition of noise had significant impact while degradation of spectral resolution had much less impact on accuracy and precision. Here, we specifically focus on aspects of remote LIBS instrument performance as they relate to various types of signal degradation. To assess the sensitivity of LIBS analysis to signal-to-noise ratio (SNR) and spectral resolution, the signal in each spectrum from a suite of 50 laboratory spectra of igneous rocks was variably degraded by increasing the peak widths (simulating misalignment) and decreasing the spectral amplitude (simulating decreases in SNR).

  1. A NASA Climate Model Data Services (CDS) End-to-End System to Support Reanalysis Intercomparison

    NASA Astrophysics Data System (ADS)

    Carriere, L.; Potter, G. L.; McInerney, M.; Nadeau, D.; Shen, Y.; Duffy, D.; Schnase, J. L.; Maxwell, T. P.; Huffer, E.

    2014-12-01

    The NASA Climate Model Data Service (CDS) and the NASA Center for Climate Simulation (NCCS) are collaborating to provide an end-to-end system for the comparative study of the major Reanalysis projects, currently, ECMWF ERA-Interim, NASA/GMAO MERRA, NOAA/NCEP CFSR, NOAA/ESRL 20CR, and JMA JRA25. Components of the system include the full spectrum of Climate Model Data Services; Data, Compute Services, Data Services, Analytic Services and Knowledge Services. The Data includes standard Reanalysis model output, and will be expanded to include gridded observations, and gridded Innovations (O-A and O-F). The NCCS High Performance Science Cloud provides the compute environment (storage, servers, and network). Data Services are provided through an Earth System Grid Federation (ESGF) data node complete with Live Access Server (LAS), Web Map Service (WMS) and Ultrascale Visualization Climate Data Analysis Tools (UV-CDAT) for visualization, as well as a collaborative interface through the Earth System CoG. Analytic Services include UV-CDAT for analysis and MERRA/AS, accessed via the CDS API, for computation services, both part of the CDS Climate Analytics as a Service (CAaaS). Knowledge Services include access to an Ontology browser, ODISEES, for metadata search and data retrieval. The result is a system that provides the ability for both reanalysis scientists and those scientists in need of reanalysis output to identify the data of interest, compare, compute, visualize, and research without the need for transferring large volumes of data, performing time consuming format conversions, and writing code for frequently run computations and visualizations.

  2. End-to-End Self-Assembly of RADA 16-I Nanofibrils in Aqueous Solutions

    PubMed Central

    Arosio, Paolo; Owczarz, Marta; Wu, Hua; Butté, Alessandro; Morbidelli, Massimo

    2012-01-01

    RADARADARADARADA (RADA 16-I) is a synthetic amphiphilic peptide designed to self-assemble in a controlled way into fibrils and higher ordered structures depending on pH. In this work, we use various techniques to investigate the state of the peptide dispersed in water under dilute conditions at different pH and in the presence of trifluoroacetic acid or hydrochloric acid. We have identified stable RADA 16-I fibrils at pH 2.0–4.5, which have a length of ∼200–400 nm and diameter of 10 nm. The fibrils have the characteristic antiparallel β-sheet structure of amyloid fibrils, as measured by circular dichroism and Fourier transform infrared spectrometry. During incubation at pH 2.0–4.5, the fibrils elongate very slowly via an end-to-end fibril-fibril aggregation mechanism, without changing their diameter, and the kinetics of such aggregation depends on pH and anion type. At pH 2.0, we also observed a substantial amount of monomers in the system, which do not participate in the fibril elongation and degrade to fragments. The fibril-fibril elongation kinetics has been simulated using the Smoluchowski kinetic model, population balance equations, and the simulation results are in good agreement with the experimental data. It is also found that the aggregation process is not limited by diffusion but rather is an activated process with energy barrier in the order of 20 kcal/mol. PMID:22500762

  3. End-to-end self-assembly of RADA 16-I nanofibrils in aqueous solutions.

    PubMed

    Arosio, Paolo; Owczarz, Marta; Wu, Hua; Butté, Alessandro; Morbidelli, Massimo

    2012-04-04

    RADARADARADARADA (RADA 16-I) is a synthetic amphiphilic peptide designed to self-assemble in a controlled way into fibrils and higher ordered structures depending on pH. In this work, we use various techniques to investigate the state of the peptide dispersed in water under dilute conditions at different pH and in the presence of trifluoroacetic acid or hydrochloric acid. We have identified stable RADA 16-I fibrils at pH 2.0-4.5, which have a length of ∼200-400 nm and diameter of 10 nm. The fibrils have the characteristic antiparallel β-sheet structure of amyloid fibrils, as measured by circular dichroism and Fourier transform infrared spectrometry. During incubation at pH 2.0-4.5, the fibrils elongate very slowly via an end-to-end fibril-fibril aggregation mechanism, without changing their diameter, and the kinetics of such aggregation depends on pH and anion type. At pH 2.0, we also observed a substantial amount of monomers in the system, which do not participate in the fibril elongation and degrade to fragments. The fibril-fibril elongation kinetics has been simulated using the Smoluchowski kinetic model, population balance equations, and the simulation results are in good agreement with the experimental data. It is also found that the aggregation process is not limited by diffusion but rather is an activated process with energy barrier in the order of 20 kcal/mol.

  4. An optimized end-to-end process for the analysis of agile earth observation satellite missions

    NASA Astrophysics Data System (ADS)

    Hahn, M.; Müller, T.; Levenhagen, J.

    2014-12-01

    Agile earth observation satellite missions are becoming more and more important due to their capability to perform fast reorientation maneuvers with 3 degrees of freedom to capture different target areas along the orbital path, thus increasing the observed area and complexity of scans. The design of an agile earth observation satellite mission is a non-trivial task due to the fact that a trade-off between observed area and complexity of the scans on the one hand and degree of agility available and thus performance of the attitude control devices on the other hand has to be done. Additionally, the designed mission has to be evaluated in a realistic environment also taking into account the specific characteristics of the chosen actuators. In the present work, several methods are combined to provide an integrated analysis of agile earth observation satellite missions starting from the definition of a desired ground scan scenario, going via the creation of a guidance profile to a realistic simulation and ending at the verification of the feasibility by detailed closed-loop simulation. Regarding its technical implementation at Astrium GmbH, well-proven tools for the different tasks of the analysis are incorporated and well defined interfaces for those tools are specified, allowing a high degree of automatism and thus saving time and minimizing errors. This results in a complete end-to-end process for the design, analysis and verification of agile earth observation satellite missions. This process is demonstrated by means of an example analysis using control moment gyros for a high agility mission.

  5. A Mechanistic End-to-End Concussion Model That Translates Head Kinematics to Neurologic Injury

    PubMed Central

    Ng, Laurel J.; Volman, Vladislav; Gibbons, Melissa M.; Phohomsiri, Pi; Cui, Jianxia; Swenson, Darrell J.; Stuhmiller, James H.

    2017-01-01

    Past concussion studies have focused on understanding the injury processes occurring on discrete length scales (e.g., tissue-level stresses and strains, cell-level stresses and strains, or injury-induced cellular pathology). A comprehensive approach that connects all length scales and relates measurable macroscopic parameters to neurological outcomes is the first step toward rationally unraveling the complexity of this multi-scale system, for better guidance of future research. This paper describes the development of the first quantitative end-to-end (E2E) multi-scale model that links gross head motion to neurological injury by integrating fundamental elements of tissue and cellular mechanical response with axonal dysfunction. The model quantifies axonal stretch (i.e., tension) injury in the corpus callosum, with axonal functionality parameterized in terms of axonal signaling. An internal injury correlate is obtained by calculating a neurological injury measure (the average reduction in the axonal signal amplitude) over the corpus callosum. By using a neurologically based quantity rather than externally measured head kinematics, the E2E model is able to unify concussion data across a range of exposure conditions and species with greater sensitivity and specificity than correlates based on external measures. In addition, this model quantitatively links injury of the corpus callosum to observed specific neurobehavioral outcomes that reflect clinical measures of mild traumatic brain injury. This comprehensive modeling framework provides a basis for the systematic improvement and expansion of this mechanistic-based understanding, including widening the range of neurological injury estimation, improving concussion risk correlates, guiding the design of protective equipment, and setting safety standards. PMID:28663736

  6. End-To-END Performance of the Future MOMA Instrument Aboard the ExoMars Mission

    NASA Astrophysics Data System (ADS)

    Pinnick, V. T.; Buch, A.; Szopa, C.; Grand, N.; Danell, R.; Grubisic, A.; van Amerom, F. H. W.; Glavin, D. P.; Freissinet, C.; Coll, P. J.; Stalport, F.; Humeau, O.; Arevalo, R. D., Jr.; Brinckerhoff, W. B.; Steininger, H.; Goesmann, F.; Raulin, F.; Mahaffy, P. R.

    2015-12-01

    Following the SAM experiment aboard the Curiosity rover, the Mars Organic Molecule Analyzer (MOMA) experiment aboard the 2018 ExoMars mission will be the continuation of the search for organic matter on the Mars surface. One advancement with the ExoMars mission is that the sample will be extracted as deep as 2 meters below the Martian surface to minimize effects of radiation and oxidation on organic materials. To analyze the wide range of organic composition (volatile and non-volatile compounds) of the Martian soil, MOMA is equipped with a dual ion source ion trap mass spectrometer utilizing UV laser desorption / ionization (LDI) and pyrolysis gas chromatography (pyr-GC). In order to analyze refractory organic compounds and chiral molecules during GC-ITMS analysis, samples may be submitted to a derivatization process, consisting of the reaction of the sample components with specific reactants (MTBSTFA [1], DMF-DMA [2] or TMAH [3]). Previous experimental reports have focused on coupling campaigns between the breadboard versions of the GC, provided by the French team (LISA, LATMOS, CentraleSupelec), and the MS, provided by the US team (NASA-GSFC). This work focuses on the performance verification and optimization of the GC-ITMS experiment using the Engineering Test Unit (ETU) models which are representative of the form, fit and function of the flight instrument including a flight-like pyrolysis oven and tapping station providing by the German team (MPS). The results obtained demonstrate the current status of the end-to-end performance of the gas chromatography-mass spectrometry mode of operation. References: [1] Buch, A. et al. (2009) J Chrom. A, 43, 143-151. [2] Freissinet et al. (2011) J Chrom A, 1306, 59-71. [3] Geffroy-Rodier, C. et al. (2009) JAAP, 85, 454-459.

  7. Advanced End-to-end Simulation for On-board Processing (AESOP)

    NASA Technical Reports Server (NTRS)

    Mazer, Alan S.

    1994-01-01

    Developers of data compression algorithms typically use their own software together with commercial packages to implement, evaluate and demonstrate their work. While convenient for an individual developer, this approach makes it difficult to build on or use another's work without intimate knowledge of each component. When several people or groups work on different parts of the same problem, the larger view can be lost. What's needed is a simple piece of software to stand in the gap and link together the efforts of different people, enabling them to build on each other's work, and providing a base for engineers and scientists to evaluate the parts as a cohesive whole and make design decisions. AESOP (Advanced End-to-end Simulation for On-board Processing) attempts to meet this need by providing a graphical interface to a developer-selected set of algorithms, interfacing with compiled code and standalone programs, as well as procedures written in the IDL and PV-Wave command languages. As a proof of concept, AESOP is outfitted with several data compression algorithms integrating previous work on different processors (AT&T DSP32C, TI TMS320C30, SPARC). The user can specify at run-time the processor on which individual parts of the compression should run. Compressed data is then fed through simulated transmission and uncompression to evaluate the effects of compression parameters, noise and error correction algorithms. The following sections describe AESOP in detail. Section 2 describes fundamental goals for usability. Section 3 describes the implementation. Sections 4 through 5 describe how to add new functionality to the system and present the existing data compression algorithms. Sections 6 and 7 discuss portability and future work.

  8. Weaving marine food webs from end to end under global change

    NASA Astrophysics Data System (ADS)

    Moloney, Coleen L.; St John, Michael A.; Denman, Kenneth L.; Karl, David M.; Köster, Friedrich W.; Sundby, Svein; Wilson, Rory P.

    2011-02-01

    Marine food web dynamics are determined by interactions within and between species and between species and their environment. Global change directly affects abiotic conditions and living organisms, impinging on all trophic levels in food webs. Different groups of marine researchers traditionally study different aspects of these changes. However, over medium to long time scales perturbations affecting food webs need to be considered across the full range from nutrients to top predators. Studies of end-to-end marine food webs not only span organism sizes and trophic levels, but should also help align multidisciplinary research to common goals and perspectives. Topics are described that bridge disciplinary gaps and are needed to develop new understanding of the reciprocal impacts of global change on marine food webs and ocean biogeochemistry. These include (1) the effects of nutrients on biomass and production, (2) the effects of varying element ratios on food web structure and food quality, (3) bulk flows of energy and material in food webs and their efficiencies of transfer, (4) the ecological effects of species richness and the roles of microbial organisms, (5) the role of feeding behaviour in food web dynamics and trophic controls, (6) the spatial dynamics of communities and links between different food webs, (7) the combined effects of body size and behaviour in determining dynamics of food webs, and (8) the extent to which the ability of marine organisms (and communities) to adapt will influence food web dynamics. An overriding issue that influences all topics concerns the time and space scales of ecosystem variability. Threads link different nodes of information among various topics, emphasizing the importance of tackling food web studies with a variety of modelling approaches and through a combination of field and experimental studies with a strong comparative approach.

  9. In vivo laser assisted end-to-end anastomosis with ICG-infused chitosan patches

    NASA Astrophysics Data System (ADS)

    Rossi, Francesca; Matteini, Paolo; Esposito, Giuseppe; Scerrati, Alba; Albanese, Alessio; Puca, Alfredo; Maira, Giulio; Rossi, Giacomo; Pini, Roberto

    2011-07-01

    Laser assisted vascular repair is a new optimized technique based on the use of ICG-infused chitosan patch to close a vessel wound, with or even without few supporting single stitches. We present an in vivo experimental study on an innovative end-to-end laser assisted vascular anastomotic (LAVA) technique, performed with the application of ICGinfused chitosan patches. The photostability and the mechanical properties of ICG-infused chitosan films were preliminary measured. The in vivo study was performed in 10 New Zealand rabbits. After anesthesia, a 3-cm segment of the right common carotid artery was exposed, thus clamped proximally and distally. The artery was then interrupted by means of a full thickness cut. Three single microsutures were used to approximate the two vessel edges. The ICG-infused chitosan patch was rolled all over the anastomotic site and welded by the use of a diode laser emitting at 810 nm and equipped with a 300 μm diameter optical fiber. Welding was obtained by delivering single laser spots to induce local patch/tissue adhesion. The result was an immediate closure of the anastomosis, with no bleeding at clamps release. Thus animals underwent different follow-up periods, in order to evaluate the welded vessels over time. At follow-up examinations, all the anastomoses were patent and no bleeding signs were documented. Samples of welded vessels underwent histological examinations. Results showed that this technique offer several advantages over conventional suturing methods: simplification of the surgical procedure, shortening of the operative time, better re-endothelization and optimal vascular healing process.

  10. Designing an End-to-End System for Data Storage, Analysis, and Visualization for an Urban Environmental Observatory

    NASA Astrophysics Data System (ADS)

    McGuire, M. P.; Welty, C.; Gangopadhyay, A.; Karabatis, G.; Chen, Z.

    2006-05-01

    The urban environment is formed by complex interactions between natural and human dominated systems, the study of which requires the collection and analysis of very large datasets that span many disciplines. Recent advances in sensor technology and automated data collection have improved the ability to monitor urban environmental systems and are making the idea of an urban environmental observatory a reality. This in turn has created a number of potential challenges in data management and analysis. We present the design of an end-to-end system to store, analyze, and visualize data from a prototype urban environmental observatory based at the Baltimore Ecosystem Study, a National Science Foundation Long Term Ecological Research site (BES LTER). We first present an object-relational design of an operational database to store high resolution spatial datasets as well as data from sensor networks, archived data from the BES LTER, data from external sources such as USGS NWIS, EPA Storet, and metadata. The second component of the system design includes a spatiotemporal data warehouse consisting of a data staging plan and a multidimensional data model designed for the spatiotemporal analysis of monitoring data. The system design also includes applications for multi-resolution exploratory data analysis, multi-resolution data mining, and spatiotemporal visualization based on the spatiotemporal data warehouse. Also the system design includes interfaces with water quality models such as HSPF, SWMM, and SWAT, and applications for real-time sensor network visualization, data discovery, data download, QA/QC, and backup and recovery, all of which are based on the operational database. The system design includes both internet and workstation-based interfaces. Finally we present the design of a laboratory for spatiotemporal analysis and visualization as well as real-time monitoring of the sensor network.

  11. Identifying Elusive Electromagnetic Counterparts to Gravitational Wave Mergers: An End-to-end Simulation

    NASA Astrophysics Data System (ADS)

    Nissanke, Samaya; Kasliwal, Mansi; Georgieva, Alexandra

    2013-04-01

    Combined gravitational wave (GW) and electromagnetic (EM) observations of compact binary mergers should enable detailed studies of astrophysical processes in the strong-field gravity regime. This decade, ground-based GW interferometers promise to routinely detect compact binary mergers. Unfortunately, networks of GW interferometers have poor angular resolution on the sky and their EM signatures are predicted to be faint. Therefore, a challenging goal will be to unambiguously pinpoint the EM counterparts of GW mergers. We perform the first comprehensive end-to-end simulation that focuses on: (1) GW sky localization, distance measures, and volume errors with two compact binary populations and four different GW networks; (2) subsequent EM detectability by a slew of multiwavelength telescopes; and (3) final identification of the merger counterpart amidst a sea of possible astrophysical false positives. First, we find that double neutron star binary mergers can be detected out to a maximum distance of 400 Mpc (or 750 Mpc) by three (or five) detector GW networks, respectively. Neutron-star-black-hole binary mergers can be detected a factor of 1.5 further out; their median to maximum sky localizations are 50-170 deg2 (or 6-65 deg2) for a three (or five) detector GW network. Second, by optimizing depth, cadence, and sky area, we quantify relative fractions of optical counterparts that are detectable by a suite of different aperture-size telescopes across the globe. Third, we present five case studies to illustrate the diversity of scenarios in secure identification of the EM counterpart. We discuss the case of a typical binary, neither beamed nor nearby, and the challenges associated with identifying an EM counterpart at both low and high Galactic latitudes. For the first time, we demonstrate how construction of low-latency GW volumes in conjunction with local universe galaxy catalogs can help solve the problem of false positives. We conclude with strategies that would

  12. An End-to-End System to Enable Quick, Easy and Inexpensive Deployment of Hydrometeorological Stations

    NASA Astrophysics Data System (ADS)

    Celicourt, P.; Piasecki, M.

    2014-12-01

    The high cost of hydro-meteorological data acquisition, communication and publication systems along with limited qualified human resources is considered as the main reason why hydro-meteorological data collection remains a challenge especially in developing countries. Despite significant advances in sensor network technologies which gave birth to open hardware and software, low-cost (less than $50) and low-power (in the order of a few miliWatts) sensor platforms in the last two decades, sensors and sensor network deployment remains a labor-intensive, time consuming, cumbersome, and thus expensive task. These factors give rise for the need to develop a affordable, simple to deploy, scalable and self-organizing end-to-end (from sensor to publication) system suitable for deployment in such countries. The design of the envisioned system will consist of a few Sensed-And-Programmed Arduino-based sensor nodes with low-cost sensors measuring parameters relevant to hydrological processes and a Raspberry Pi micro-computer hosting the in-the-field back-end data management. This latter comprises the Python/Django model of the CUAHSI Observations Data Model (ODM) namely DjangODM backed by a PostgreSQL Database Server. We are also developing a Python-based data processing script which will be paired with the data autoloading capability of Django to populate the DjangODM database with the incoming data. To publish the data, the WOFpy (WaterOneFlow Web Services in Python) developed by the Texas Water Development Board for 'Water Data for Texas' which can produce WaterML web services from a variety of back-end database installations such as SQLite, MySQL, and PostgreSQL will be used. A step further would be the development of an appealing online visualization tool using Python statistics and analytics tools (Scipy, Numpy, Pandas) showing the spatial distribution of variables across an entire watershed as a time variant layer on top of a basemap.

  13. Astra: Interdisciplinary study on enhancement of the end-to-end accuracy for spacecraft tracking techniques

    NASA Astrophysics Data System (ADS)

    Iess, Luciano; Di Benedetto, Mauro; James, Nick; Mercolino, Mattia; Simone, Lorenzo; Tortora, Paolo

    2014-02-01

    Navigation of deep-space probes is accomplished through a variety of different radio observables, namely Doppler, ranging and Delta-Differential One-Way Ranging (Delta-DOR). The particular mix of observations used for navigation mainly depends on the available on-board radio system, the mission phase and orbit determination requirements. The accuracy of current ESA and NASA tracking systems is at level of 0.1 mm/s at 60 s integration time for Doppler, 1-5 m for ranging and 6-15 nrad for Delta-DOR measurements in a wide range of operational conditions. The ASTRA study, funded under ESA's General Studies Programme (GSP), addresses the ways to improve the end-to-end accuracy of Doppler, ranging and Delta-DOR systems by roughly a factor of 10. The target accuracies were set to 0.01 mm/s at 60 s integration time for Doppler, 20 cm for ranging and 1 nrad for Delta-DOR. The companies and universities that took part in the study were the University of Rome Sapienza, ALMASpace, BAE Systems and Thales Alenia Space Italy. The analysis of an extensive data set of radio-metric observables and dedicated tests of the ground station allowed consolidating the error budget for each measurement technique. The radio-metric data set comprises X/X, X/Ka and Ka/Ka range and Doppler observables from the Cassini and Rosetta missions. It includes also measurements from the Advanced Media Calibration System (AMCS) developed by JPL for the radio science experiments of the Cassini mission. The error budget for the three radio-metric observables was consolidated by comparing the statistical properties of the data set with the expected error models. The analysis confirmed the contribution from some error sources, but revealed also some discrepancies and ultimately led to improved error models. The error budget reassessment provides adequate information for building guidelines and strategies to effectively improve the navigation accuracies of future deep space missions. We report both on updated

  14. End-To-END Performance of the future MOMA intrument aboard the EXOMARS MISSION

    NASA Astrophysics Data System (ADS)

    Buch, A.; Pinnick, V. T.; Szopa, C.; Grand, N.; Danell, R.; van Amerom, F. H. W.; Freissinet, C.; Glavin, D. P.; Stalport, F.; Arevalo, R. D., Jr.; Coll, P. J.; Steininger, H.; Raulin, F.; Goesmann, F.; Mahaffy, P. R.; Brinckerhoff, W. B.

    2016-12-01

    After the SAM experiment aboard the curiosity rover, the Mars Organic Molecule Analyzer (MOMA) experiment aboard the future ExoMars mission will be the continuation of the search for the organic composition of the Mars surface with the advantage that the sample will be extracted as deep as 2 meters below the martian surface to minimize effects of radiation and oxidation on organic materials. To analyse the wide range of organic composition (volatile and non volatils compounds) of the martian soil MOMA is composed with an UV laser desorption / ionization (LDI) and a pyrolysis gas chromatography ion trap mass spectrometry (pyr-GC-ITMS). In order to analyse refractory organic compounds and chirality samples which undergo GC-ITMS analysis may be submitted to a derivatization process, consisting of the reaction of the sample components with specific reactants (MTBSTFA [1], DMF-DMA [2] or TMAH [3]). To optimize and test the performance of the GC-ITMS instrument we have performed several coupling tests campaigns between the GC, providing by the French team (LISA, LATMOS, CentraleSupelec), and the MS, providing by the US team (NASA, GSFC). Last campaign has been done with the ITU models wich is similar to the flight model and wich include the oven and the taping station providing by the German team (MPS). The results obtained demonstrate the current status of the end-to-end performance of the gas chromatography-mass spectrometry mode of operation. References:[1] Buch, A. et al. (2009) J chrom. A, 43, 143-151. [2] Freissinet et al. (2011) J Chrom A, 1306, 59-71. [3] Geffroy-Rodier, C. et al. (2009) JAAP, 85, 454-459. Acknowledgements: Funding provided by the Mars Exploration Program (point of contact, George Tahu, NASA/HQ). MOMA is a collaboration between NASA and ESA (PI Goesmann, MPS). MOMA-GC team acknowledges support from the French Space Agency (CNES), French National Programme of Planetology (PNP), National French Council (CNRS), Pierre Simon Laplace Institute.

  15. Urban Biomining Meets Printable Electronics: End-To-End at Destination Biological Recycling and Reprinting

    NASA Technical Reports Server (NTRS)

    Rothschild, Lynn J. (Principal Investigator); Koehne, Jessica; Gandhiraman, Ram; Navarrete, Jesica; Spangle, Dylan

    2017-01-01

    Space missions rely utterly on metallic components, from the spacecraft to electronics. Yet, metals add mass, and electronics have the additional problem of a limited lifespan. Thus, current mission architectures must compensate for replacement. In space, spent electronics are discarded; on earth, there is some recycling but current processes are toxic and environmentally hazardous. Imagine instead an end-to-end recycling of spent electronics at low mass, low cost, room temperature, and in a non-toxic manner. Here, we propose a solution that will not only enhance mission success by decreasing upmass and providing a fresh supply of electronics, but in addition has immediate applications to a serious environmental issue on the Earth. Spent electronics will be used as feedstock to make fresh electronic components, a process we will accomplish with so-called 'urban biomining' using synthetically enhanced microbes to bind metals with elemental specificity. To create new electronics, the microbes will be used as 'bioink' to print a new IC chip, using plasma jet electronics printing. The plasma jet electronics printing technology will have the potential to use martian atmospheric gas to print and to tailor the electronic and chemical properties of the materials. Our preliminary results have suggested that this process also serves as a purification step to enhance the proportion of metals in the 'bioink'. The presence of electric field and plasma can ensure printing in microgravity environment while also providing material morphology and electronic structure tunabiity and thus optimization. Here we propose to increase the TRL level of the concept by engineering microbes to dissolve the siliceous matrix in the IC, extract copper from a mixture of metals, and use the microbes as feedstock to print interconnects using mars gas simulant. To assess the ability of this concept to influence mission architecture, we will do an analysis of the infrastructure required to execute

  16. IDENTIFYING ELUSIVE ELECTROMAGNETIC COUNTERPARTS TO GRAVITATIONAL WAVE MERGERS: AN END-TO-END SIMULATION

    SciTech Connect

    Nissanke, Samaya; Georgieva, Alexandra; Kasliwal, Mansi

    2013-04-20

    Combined gravitational wave (GW) and electromagnetic (EM) observations of compact binary mergers should enable detailed studies of astrophysical processes in the strong-field gravity regime. This decade, ground-based GW interferometers promise to routinely detect compact binary mergers. Unfortunately, networks of GW interferometers have poor angular resolution on the sky and their EM signatures are predicted to be faint. Therefore, a challenging goal will be to unambiguously pinpoint the EM counterparts of GW mergers. We perform the first comprehensive end-to-end simulation that focuses on: (1) GW sky localization, distance measures, and volume errors with two compact binary populations and four different GW networks; (2) subsequent EM detectability by a slew of multiwavelength telescopes; and (3) final identification of the merger counterpart amidst a sea of possible astrophysical false positives. First, we find that double neutron star binary mergers can be detected out to a maximum distance of 400 Mpc (or 750 Mpc) by three (or five) detector GW networks, respectively. Neutron-star-black-hole binary mergers can be detected a factor of 1.5 further out; their median to maximum sky localizations are 50-170 deg{sup 2} (or 6-65 deg{sup 2}) for a three (or five) detector GW network. Second, by optimizing depth, cadence, and sky area, we quantify relative fractions of optical counterparts that are detectable by a suite of different aperture-size telescopes across the globe. Third, we present five case studies to illustrate the diversity of scenarios in secure identification of the EM counterpart. We discuss the case of a typical binary, neither beamed nor nearby, and the challenges associated with identifying an EM counterpart at both low and high Galactic latitudes. For the first time, we demonstrate how construction of low-latency GW volumes in conjunction with local universe galaxy catalogs can help solve the problem of false positives. We conclude with strategies

  17. Achieving End-to-End QoS in the Next Generation Internet: Integrated Services over Differentiated Service Networks

    NASA Technical Reports Server (NTRS)

    Bai, Haowei; Atiquzzaman, Mohammed; Ivancic, William

    2001-01-01

    Currently there are two approaches to provide Quality of Service (QoS) in the next generation Internet: An early one is the Integrated Services (IntServ) with the goal of allowing end-to-end QoS to be provided to applications; the other one is the Differentiated Services (DiffServ) architecture providing QoS in the backbone. In this context, a DiffServ network may be viewed as a network element in the total end-to-end path. The objective of this paper is to investigate the possibility of providing end-to-end QoS when IntServ runs over DiffServ backbone in the next generation Internet. Our results show that the QoS requirements of IntServ applications can be successfully achieved when IntServ traffic is mapped to the DiffServ domain in next generation Internet.

  18. Achieving End-to-End QoS in the Next Generation Internet: Integrated Services Over Differentiated Service Networks

    NASA Technical Reports Server (NTRS)

    Bai, Haowei; Atiquzzaman, Mohammed; Ivancic, William

    2001-01-01

    Currently there are two approaches to provide Quality of Service (QoS) in the next generation Internet: An early one is the Integrated Services (IntServ) with the goal of allowing end-to-end QoS to be provided to applications; the other one is the Differentiated Services (DiffServ) architecture providing QoS in the backbone. In this context, a DiffServ network may be viewed as a network element in the total end-to-end path. The objective of this paper is to investigate the possibility of providing end-to-end QoS when IntServ runs over DiffServ backbone in the next generation Internet. Our results show that the QoS requirements of IntServ applications can be successfully achieved when IntServ traffic is mapped to the DiffServ domain in next generation Internet.

  19. Unidata's Vision for Providing Comprehensive and End-to-end Data Services

    NASA Astrophysics Data System (ADS)

    Ramamurthy, M. K.

    2009-05-01

    This paper presents Unidata's vision for providing comprehensive, well-integrated, and end-to-end data services for the geosciences. These include an array of functions for collecting, finding, and accessing data; data management tools for generating, cataloging, and exchanging metadata; and submitting or publishing, sharing, analyzing, visualizing, and integrating data. When this vision is realized, users no matter where they are or how they are connected to the Internetwill be able to find and access a plethora of geosciences data and use Unidata-provided tools and services both productively and creatively in their research and education. What that vision means for the Unidata community is elucidated by drawing a simple analogy. Most of users are familiar with Amazon and eBay e-commerce sites and content sharing sites like YouTube and Flickr. On the eBay marketplace, people can sell practically anything at any time and buyers can share their experience of purchasing a product or the reputation of a seller. Likewise, at Amazon, thousands of merchants sell their goods and millions of customers not only buy those goods, but provide a review or opinion of the products they buy and share their experiences as purchasers. Similarly, YouTube and Flickr are sites tailored to video- and photo-sharing, respectively, where users can upload their own content and share it with millions of other users, including family and friends. What all these sites, together with social-networking applications like MySpace and Facebook, have enabled is a sense of a virtual community in which users can search and browse products or content, comment and rate those products from anywhere, at any time, and via any Internet- enabled device like an iPhone, laptop, or a desktop computer. In essence, these enterprises have fundamentally altered people's buying modes and behavior toward purchases. Unidata believes that similar approaches, appropriately tailored to meet the needs of the scientific

  20. Common Patterns with End-to-end Interoperability for Data Access

    NASA Astrophysics Data System (ADS)

    Gallagher, J.; Potter, N.; Jones, M. B.

    2010-12-01

    At first glance, using common storage formats and open standards should be enough to ensure interoperability between data servers and client applications, but that is often not the case. In the REAP (Realtime Environment for Analytical Processing; NSF #0619060) project we integrated access to data from OPeNDAP servers into the Kepler workflow system and found that, as in previous cases, we spent the bulk of our effort addressing the twin issues of data model compatibility and integration strategies. Implementing seamless data access between a remote data source and a client application (data sink) can be broken down into two kinds of issues. First, the solution must address any differences in the data models used by the data source (OPeNDAP) and the data sink (the Kepler workflow system). If these models match completely, there is little work to be done. However, that is rarely the case. To map OPeNDAP's data model to Kepler's, we used two techniques (ignoring trivial conversions): On-the-fly type mapping and out-of-band communication. Type conversion takes place both for data and metadata because Kepler requires a priori knowledge of some aspects (e.g., syntactic metadata) of the data to build a workflow. In addition, OPeNDAP's constraint expression syntax was used to send out-of-band information to restrict the data requested from the server, facilitating changes in the returned data's type. This technique provides a way for users to exert fine-grained control over the data request, a potentially useful technique, at the cost of requiring that users understand a little about the data source's processing capabilities. The second set of issues for end-to-end data access are integration strategies. OPeNDAP provides several different tools for bringing data into an application: C++, C and Java libraries that provide functions for newly written software; The netCDF library which enables existing applications to read from servers using an older interface; and simple

  1. Results from Solar Reflective Band End-to-End Testing for VIIRS F1 Sensor Using T-SIRCUS

    NASA Technical Reports Server (NTRS)

    McIntire, Jeff; Moyer, David; McCarthy, James K.; DeLuccia, Frank; Xiong, Xiaoxiong; Butler, James J.; Guenther, Bruce

    2011-01-01

    Verification of the Visible Infrared Imager Radiometer Suite (VIIRS) End-to-End (E2E) sensor calibration is highly recommended before launch, to identify any anomalies and to improve our understanding of the sensor on-orbit calibration performance. E2E testing of the Reflective Solar Bands (RSB) calibration cycle was performed pre-launch for the VIIRS Fight 1 (F1) sensor at the Ball Aerospace facility in Boulder CO in March 2010. VIIRS reflective band calibration cycle is very similar to heritage sensor MODIS in that solar illumination, via a diffuser, is used to correct for temporal variations in the instrument responsivity. Monochromatic light from the NIST T-SIRCUS was used to illuminate both the Earth View (EV), via an integrating sphere, and the Solar Diffuser (SD) view, through a collimator. The collimator illumination was cycled through a series of angles intended to simulate the range of possible angles for which solar radiation will be incident on the solar attenuation screen on-orbit. Ideally, the measured instrument responsivity (defined here as the ratio of the detector response to the at-sensor radiance) should be the same whether the EV or SD view is illuminated. The ratio of the measured responsivities was determined at each collimator angle and wavelength. In addition, the Solar Diffuser Stability Monitor (SDSM), a ratioing radiometer designed to track the temporal variation in the SD BRF by direct comparison to solar radiation, was illuminated by the collimator. The measured SDSM ratio was compared to the predicted ratio. An uncertainty analysis was also performed on both the SD and SDSM calibrations.

  2. Dynamic Hop Service Differentiation Model for End-to-End QoS Provisioning in Multi-Hop Wireless Networks

    NASA Astrophysics Data System (ADS)

    Youn, Joo-Sang; Seok, Seung-Joon; Kang, Chul-Hee

    This paper presents a new QoS model for end-to-end service provisioning in multi-hop wireless networks. In legacy IEEE 802.11e based multi-hop wireless networks, the fixed assignment of service classes according to flow's priority at every node causes priority inversion problem when performing end-to-end service differentiation. Thus, this paper proposes a new QoS provisioning model called Dynamic Hop Service Differentiation (DHSD) to alleviate the problem and support effective service differentiation between end-to-end nodes. Many previous works for QoS model through the 802.11e based service differentiation focus on packet scheduling on several service queues with different service rate and service priority. Our model, however, concentrates on a dynamic class selection scheme, called Per Hop Class Assignment (PHCA), in the node's MAC layer, which selects a proper service class for each packet, in accordance with queue states and service requirement, in every node along the end-to-end route of the packet. The proposed QoS solution is evaluated using the OPNET simulator. The simulation results show that the proposed model outperforms both best-effort and 802.11e based strict priority service models in mobile ad hoc environments.

  3. A vision for end-to-end data services to foster international partnerships through data sharing

    NASA Astrophysics Data System (ADS)

    Ramamurthy, M.; Yoksas, T.

    2009-04-01

    Increasingly, the conduct of science requires scientific partnerships and sharing of knowledge, information, and other assets. This is particularly true in our field where the highly-coupled Earth system and its many linkages have heightened the importance of collaborations across geographic, disciplinary, and organizational boundaries. The climate system, for example, is far too complex a puzzle to be unraveled by individual investigators or nations. As articulated in the NSF Strategic Plan: FY 2006-2011, "…discovery increasingly requires expertise of individuals from different disciplines, with diverse perspectives, and often from different nations, working together to accommodate the extraordinary complexity of today's science and engineering challenges." The Nobel Prize winning IPCC assessments are a prime example of such an effort. Earth science education is also uniquely suited to drawing connections between the dynamic Earth system and societal issues. Events like the 2004 Indian Ocean tsunami and Hurricane Katrina provide ample evidence of this relevance, as they underscore the importance of timely and interdisciplinary integration and synthesis of data. Our success in addressing such complex problems and advancing geosciences depends on the availability of a state-of-the-art and robust cyberinfrastructure, transparent and timely access to high-quality data from diverse sources, and requisite tools to integrate and use the data effectively, toward creating new knowledge. To that end, Unidata's vision calls for providing comprehensive, well-integrated, and end-to-end data services for the geosciences. These include an array of functions for collecting, finding, and accessing data; data management tools for generating, cataloging, and exchanging metadata; and submitting or publishing, sharing, analyzing, visualizing, and integrating data. When this vision is realized, users — no matter where they are, how they are connected to the Internet, or what

  4. End-to-end average BER analysis for multihop free-space optical communications with pointing errors

    NASA Astrophysics Data System (ADS)

    Sheng, Ming; Jiang, Peng; Hu, Qingsong; Su, Qin; Xie, Xiu-xiu

    2013-05-01

    This paper addresses the end-to-end average BER (ABER) performance of decode-and-forward (DF) relay free-space optical (FSO) communications over weak and strong turbulence channels with pointing errors. For the weak and strong turbulence channels, the probability distribution function (PDF) of the irradiance can be modeled by a lognormal and Gamma-Gamma distribution, respectively. Considering the effects from atmospheric attenuation, turbulence and pointing errors, we present a statistical model for the optical intensity fluctuation at the receiver. Then the end-to-end ABER performances are analyzed and derived closed-form expressions are obtained. The simulation results indicate that the derived closed-form expressions provide sufficiently accurate approximations.

  5. End-to-End Demonstrator of the Safe Affordable Fission Engine (SAFE) 30: Power Conversion and Ion Engine Operation

    NASA Technical Reports Server (NTRS)

    Hrbud, Ivana; VanDyke, Melissa; Houts, Mike; Goodfellow, Keith; Schafer, Charles (Technical Monitor)

    2001-01-01

    The Safe Affordable Fission Engine (SAFE) test series addresses Phase 1 Space Fission Systems issues in particular non-nuclear testing and system integration issues leading to the testing and non-nuclear demonstration of a 400-kW fully integrated flight unit. The first part of the SAFE 30 test series demonstrated operation of the simulated nuclear core and heat pipe system. Experimental data acquired in a number of different test scenarios will validate existing computational models, demonstrated system flexibility (fast start-ups, multiple start-ups/shut downs), simulate predictable failure modes and operating environments. The objective of the second part is to demonstrate an integrated propulsion system consisting of a core, conversion system and a thruster where the system converts thermal heat into jet power. This end-to-end system demonstration sets a precedent for ground testing of nuclear electric propulsion systems. The paper describes the SAFE 30 end-to-end system demonstration and its subsystems.

  6. Exploring the requirements for multimodal interaction for mobile devices in an end-to-end journey context.

    PubMed

    Krehl, Claudia; Sharples, Sarah

    2012-01-01

    The paper investigates the requirements for multimodal interaction on mobile devices in an end-to-end journey context. Traditional interfaces are deemed cumbersome and inefficient for exchanging information with the user. Multimodal interaction provides a different user-centred approach allowing for more natural and intuitive interaction between humans and computers. It is especially suitable for mobile interaction as it can overcome additional constraints including small screens, awkward keypads, and continuously changing settings - an inherent property of mobility. This paper is based on end-to-end journeys where users encounter several contexts during their journeys. Interviews and focus groups explore the requirements for multimodal interaction design for mobile devices by examining journey stages and identifying the users' information needs and sources. Findings suggest that multimodal communication is crucial when users multitask. Choosing suitable modalities depend on user context, characteristics and tasks.

  7. POST2 End-To-End Descent and Landing Simulation for the Autonomous Landing and Hazard Avoidance Technology Project

    NASA Technical Reports Server (NTRS)

    Fisher, Jody l.; Striepe, Scott A.

    2007-01-01

    The Program to Optimize Simulated Trajectories II (POST2) is used as a basis for an end-to-end descent and landing trajectory simulation that is essential in determining the design and performance capability of lunar descent and landing system models and lunar environment models for the Autonomous Landing and Hazard Avoidance Technology (ALHAT) project. This POST2-based ALHAT simulation provides descent and landing simulation capability by integrating lunar environment and lander system models (including terrain, sensor, guidance, navigation, and control models), along with the data necessary to design and operate a landing system for robotic, human, and cargo lunar-landing success. This paper presents the current and planned development and model validation of the POST2-based end-to-end trajectory simulation used for the testing, performance and evaluation of ALHAT project system and models.

  8. CUSat: An End-to-End In-Orbit Inspection System University Nanosatellite Program

    DTIC Science & Technology

    2007-01-01

    a Figure 5.5 2~h7 0 CAMtei Stea ,Roso,m TIC nee w warraoy- w It CT ina 2 *_o 15D" "w PROP u Da DSct CM- A-eA Fligh m-ep - *900 s ~. 220010* 108 usaw0...0280 (II...)jes 7 oni ត Onv_! 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR( S ) 5d. PROJECT NUMBER Mason A. Peck Se. TASK NUMBER 5f. WORK UNIT NUMBER 7...PERFORMING ORGANIZATION NAME( S ) AND ADDRESS(ES) 8. PERFORMING ORGANIZATION REPORT NUMBER Cornell University 9. SPONSORING I MONITORING AGENCY NAME( S ) AND

  9. Minimizing End-to-End Interference in I/O Stacks Spanning Shared Multi-Level Buffer Caches

    ERIC Educational Resources Information Center

    Patrick, Christina M.

    2011-01-01

    This thesis presents an end-to-end interference minimizing uniquely designed high performance I/O stack that spans multi-level shared buffer cache hierarchies accessing shared I/O servers to deliver a seamless high performance I/O stack. In this thesis, I show that I can build a superior I/O stack which minimizes the inter-application interference…

  10. Minimizing End-to-End Interference in I/O Stacks Spanning Shared Multi-Level Buffer Caches

    ERIC Educational Resources Information Center

    Patrick, Christina M.

    2011-01-01

    This thesis presents an end-to-end interference minimizing uniquely designed high performance I/O stack that spans multi-level shared buffer cache hierarchies accessing shared I/O servers to deliver a seamless high performance I/O stack. In this thesis, I show that I can build a superior I/O stack which minimizes the inter-application interference…

  11. Integrated End-to -End Radar Signal & Data Processing With Over-arching Knowledge-Based Control

    DTIC Science & Technology

    2006-09-01

    perform better if it knew where potential jammers were located and their characteristics. Capraro, G.T. (2006) Integrated End-to-End Radar Signal...this system by using fiber optic or wire links on board the platform. Radio frequency (RF) links using Bluetooth or 802.11 technologies can be...the mission. During flight, the KBC will receive information about weather, jammer locations, requests for information, discrete locations, fusion

  12. Sequence and Temperature Dependence of the End-to-End Collision Dynamics of Single-Stranded DNA

    PubMed Central

    Uzawa, Takanori; Isoshima, Takashi; Ito, Yoshihiro; Ishimori, Koichiro; Makarov, Dmitrii E.; Plaxco, Kevin W.

    2013-01-01

    Intramolecular collision dynamics play an essential role in biomolecular folding and function and, increasingly, in the performance of biomimetic technologies. To date, however, the quantitative studies of dynamics of single-stranded nucleic acids have been limited. Thus motivated, here we investigate the sequence composition, chain-length, viscosity, and temperature dependencies of the end-to-end collision dynamics of single-stranded DNAs. We find that both the absolute collision rate and the temperature dependencies of these dynamics are base-composition dependent, suggesting that base stacking interactions are a significant contributor. For example, whereas the end-to-end collision dynamics of poly-thymine exhibit simple, linear Arrhenius behavior, the behavior of longer poly-adenine constructs is more complicated. Specifically, 20- and 25-adenine constructs exhibit biphasic temperature dependencies, with their temperature dependences becoming effectively indistinguishable from that of poly-thymine above 335 K for 20-adenines and 328 K for 25-adenines. The differing Arrhenius behaviors of poly-thymine and poly-adenine and the chain-length dependence of the temperature at which poly-adenine crosses over to behave like poly-thymine can be explained by a barrier friction mechanism in which, at low temperatures, the energy barrier for the local rearrangement of poly-adenine becomes the dominant contributor to its end-to-end collision dynamics. PMID:23746521

  13. Treatment of distal splenic artery aneurysm by laparoscopic aneurysmectomy with end-to-end anastomosis: A case report.

    PubMed

    Zhang, Yuhua; Liu, Zhenjie; Shen, Guoliang; Zhang, Jungang; Assa, Carmel Rebecca; Hong, Defei

    2017-05-01

    Splenic artery aneurysm (SAA) is the most common visceral artery aneurysm, while most SAAs are treated by endovascular or open procedures. Here we present a case of SAA treated by laparoscopic aneurysmectomy with end-to-end anastomosis. A 40-year-old woman was incidentally found to have an asymptomatic distal SAA. CT scan revealed the SAA to be located at the hilum of the spleen, with a maximal diameter of 2.7 cm. To prevent sudden rupture, the patient received laparoscopic aneurysmectomy. During the operation, end-to-end anastomosis was also performed since a tortuous proximal splenic artery prevented delivery of the stent graft. The patient was ambulated 12 hours after surgery and discharged 5 days later. Postoperative recovery was smooth without hemorrhage, infarction, infection, or splenic artery thrombosis. At 10-month follow-up,no hemorrhage, aneurysm recurrence, spleen infarction, splenic artery stenosis, or thrombosis had occurred. Patients with distal SAA can be treated by laparoscopic aneurysmectomy with end-to-end anastomosis to preserve the spleen. The laparoscopic procedure is safe and feasible in the selected patients.

  14. Debris mitigation measures by satellite design and operational methods - Findings from the DLR space debris End-to-End Service

    NASA Astrophysics Data System (ADS)

    Sdunnus, H.; Beltrami, P.; Janovsky, R.; Koppenwallner, G.; Krag, H.; Reimerdes, H.; Schäfer, F.

    Debris Mitigation has been recognised as an issue to be addressed by the space faring nations around the world. Currently, there are various activities going on, aiming at the establishment of debris mitigation guidelines on various levels, reaching from the UN down to national space agencies. Though guidelines established on the national level already provide concrete information how things should be done (rather that specifying what should be done or providing fundamental principles) potential users of the guidelines will still have the need to explore the technical, management, and financial implications of the guidelines for their projects. Those questions are addressed by the so called "Space Debris End-to-End Service" project, which has been initiated as a national initiative of the German Aerospace Centre (DLR). Based on a review of already existing mitigation guidelines or guidelines under development and following an identification of needs from a circle of industrial users the "End-to-End Service Gu idelines" have been established for designer and operators of spacecraft. The End-to-End Service Guidelines are based on requirements addressed by the mitigation guidelines and provide recommendations how and when the technical consideration of the mitigation guidelines should take place. By referencing requirements from the mitigation guidelines, the End-to-End Service Guidelines address the consideration of debris mitigation measures by spacecraft design and operational measures. This paper will give an introduction to the End-to-End Service Guidelines. It will focus on the proposals made for mitigation measures by the S/C system design, i.e. on protective design measures inside the spacecraft and on design measures, e.g. innovative protective (shielding) systems. Furthermore, approaches on the analytical optimisation of protective systems will be presented, aiming at the minimisation of shield mass under conservation of the protective effects. On the

  15. The Kepler End-to-End Data Pipeline: From Photons to Far Away Worlds

    NASA Technical Reports Server (NTRS)

    Cooke, Brian; Thompson, Richard; Standley, Shaun

    2012-01-01

    The Kepler mission is described in overview and the Kepler technique for discovering exoplanets is discussed. The design and implementation of the Kepler spacecraft, tracing the data path from photons entering the telescope aperture through raw observation data transmitted to the ground operations team is described. The technical challenges of operating a large aperture photometer with an unprecedented 95 million pixel detector are addressed as well as the onboard technique for processing and reducing the large volume of data produced by the Kepler photometer. The technique and challenge of day-to-day mission operations that result in a very high percentage of time on target is discussed. This includes the day to day process for monitoring and managing the health of the spacecraft, the annual process for maintaining sun on the solar arrays while still keeping the telescope pointed at the fixed science target, the process for safely but rapidly returning to science operations after a spacecraft initiated safing event and the long term anomaly resolution process.The ground data processing pipeline, from the point that science data is received on the ground to the presentation of preliminary planetary candidates and supporting data to the science team for further evaluation is discussed. Ground management, control, exchange and storage of Kepler's large and growing data set is discussed as well as the process and techniques for removing noise sources and applying calibrations to intermediate data products.

  16. Exploring Two Approaches for an End-to-End Scientific Analysis Workflow

    NASA Astrophysics Data System (ADS)

    Dodelson, Scott; Kent, Steve; Kowalkowski, Jim; Paterno, Marc; Sehrish, Saba

    2015-12-01

    The scientific discovery process can be advanced by the integration of independently-developed programs run on disparate computing facilities into coherent workflows usable by scientists who are not experts in computing. For such advancement, we need a system which scientists can use to formulate analysis workflows, to integrate new components to these workflows, and to execute different components on resources that are best suited to run those components. In addition, we need to monitor the status of the workflow as components get scheduled and executed, and to access the intermediate and final output for visual exploration and analysis. Finally, it is important for scientists to be able to share their workflows with collaborators. We have explored two approaches for such an analysis framework for the Large Synoptic Survey Telescope (LSST) Dark Energy Science Collaboration (DESC); the first one is based on the use and extension of Galaxy, a web-based portal for biomedical research, and the second one is based on a programming language, Python. In this paper, we present a brief description of the two approaches, describe the kinds of extensions to the Galaxy system we have found necessary in order to support the wide variety of scientific analysis in the cosmology community, and discuss how similar efforts might be of benefit to the HEP community.

  17. Exploring Two Approaches for an End-to-End Scientific Analysis Workflow

    SciTech Connect

    Dodelson, Scott; Kent, Steve; Kowalkowski, Jim; Paterno, Marc; Sehrish, Saba

    2015-12-23

    The advance of the scientific discovery process is accomplished by the integration of independently-developed programs run on disparate computing facilities into coherent workflows usable by scientists who are not experts in computing. For such advancement, we need a system which scientists can use to formulate analysis workflows, to integrate new components to these workflows, and to execute different components on resources that are best suited to run those components. In addition, we need to monitor the status of the workflow as components get scheduled and executed, and to access the intermediate and final output for visual exploration and analysis. Finally, it is important for scientists to be able to share their workflows with collaborators. Moreover we have explored two approaches for such an analysis framework for the Large Synoptic Survey Telescope (LSST) Dark Energy Science Collaboration (DESC), the first one is based on the use and extension of Galaxy, a web-based portal for biomedical research, and the second one is based on a programming language, Python. In our paper, we present a brief description of the two approaches, describe the kinds of extensions to the Galaxy system we have found necessary in order to support the wide variety of scientific analysis in the cosmology community, and discuss how similar efforts might be of benefit to the HEP community.

  18. Exploring Two Approaches for an End-to-End Scientific Analysis Workflow

    DOE PAGES

    Dodelson, Scott; Kent, Steve; Kowalkowski, Jim; ...

    2015-12-23

    The advance of the scientific discovery process is accomplished by the integration of independently-developed programs run on disparate computing facilities into coherent workflows usable by scientists who are not experts in computing. For such advancement, we need a system which scientists can use to formulate analysis workflows, to integrate new components to these workflows, and to execute different components on resources that are best suited to run those components. In addition, we need to monitor the status of the workflow as components get scheduled and executed, and to access the intermediate and final output for visual exploration and analysis. Finally,more » it is important for scientists to be able to share their workflows with collaborators. Moreover we have explored two approaches for such an analysis framework for the Large Synoptic Survey Telescope (LSST) Dark Energy Science Collaboration (DESC), the first one is based on the use and extension of Galaxy, a web-based portal for biomedical research, and the second one is based on a programming language, Python. In our paper, we present a brief description of the two approaches, describe the kinds of extensions to the Galaxy system we have found necessary in order to support the wide variety of scientific analysis in the cosmology community, and discuss how similar efforts might be of benefit to the HEP community.« less

  19. The Kepler End-to-End Data Pipeline: From Photons to Far Away Worlds

    NASA Technical Reports Server (NTRS)

    Cooke, Brian; Thompson, Richard; Standley, Shaun

    2012-01-01

    The Kepler mission is described in overview and the Kepler technique for discovering exoplanets is discussed. The design and implementation of the Kepler spacecraft, tracing the data path from photons entering the telescope aperture through raw observation data transmitted to the ground operations team is described. The technical challenges of operating a large aperture photometer with an unprecedented 95 million pixel detector are addressed as well as the onboard technique for processing and reducing the large volume of data produced by the Kepler photometer. The technique and challenge of day-to-day mission operations that result in a very high percentage of time on target is discussed. This includes the day to day process for monitoring and managing the health of the spacecraft, the annual process for maintaining sun on the solar arrays while still keeping the telescope pointed at the fixed science target, the process for safely but rapidly returning to science operations after a spacecraft initiated safing event and the long term anomaly resolution process.The ground data processing pipeline, from the point that science data is received on the ground to the presentation of preliminary planetary candidates and supporting data to the science team for further evaluation is discussed. Ground management, control, exchange and storage of Kepler's large and growing data set is discussed as well as the process and techniques for removing noise sources and applying calibrations to intermediate data products.

  20. An End-to-End Architecture for Science Goal Driven Observing

    NASA Technical Reports Server (NTRS)

    Jones, Jeremy; Grosvenor, Sandy; Koratkar, Anuradha; Memarsadeghi, Nargess; Wolf, Karl; Obenschain, Arthur F. (Technical Monitor)

    2002-01-01

    New observatories will have greater on-board storage capacity and on-board processing capabilities. The new bottleneck will be download capacity. The cost of downlink time and limitations of bandwidth will end the era where all exposure data is downloaded and all data processing is performed on the ground. In addition, observing campaigns involving inherently variable targets will need scheduling flexibility to focus observing time and data download on exposures that are scientifically interesting. The ability to quickly recognize and react to such events by re-prioritizing the observing schedule will be an essential characteristic for maximizing scientific returns. It will also be a step towards increasing spacecraft autonomy, a major goal of NASA's strategic plan. The science goal monitoring (SGM) system is a proof-of-concept effort to address these challenges. We are developing an interactive distributed system that will use on-board processing and storage combined with event-driven interfaces with ground-based processing and operations, to enable fast re-prioritization of observing schedules, and to minimize time spent on non-optimized observations. SGM is initially aimed towards time-tagged observing modes used frequently in spectroscopic studies of varying targets. In particular, the SGM is collaborating with the proposed MIDEX-class mission Kronos team. The variable targets that Kronos seeks to study make an adaptive system such as SGM particularly valuable for achieving mission goals. However, the architecture and interfaces will also be designed for easy adaptability to other observing platforms, including ground-based systems and to work with different scheduling and pipeline processing systems. This talk will focus on our strategy for developing SGM and the technical challenges that we have encountered. We will discuss the SGM architecture as it applies to the Kronos mission and explain how it is scalable to other missions.

  1. Design to monitor trend in abundance and presence of American beaver (Castor canadensis) at the national forest scale.

    PubMed

    Beck, Jeffrey L; Dauwalter, Daniel C; Gerow, Kenneth G; Hayward, Gregory D

    2010-05-01

    Wildlife conservationists design monitoring programs to assess population dynamics, project future population states, and evaluate the impacts of management actions on populations. Because agency mandates and conservation laws call for monitoring data to elicit management responses, it is imperative to design programs that match the administrative scale for which management decisions are made. We describe a program to monitor population trends in American beaver (Castor canadensis) on the US Department of Agriculture, Black Hills National Forest (BHNF) in southwestern South Dakota and northeastern Wyoming, USA. Beaver have been designated as a management indicator species on the BHNF because of their association with riparian and aquatic habitats and its status as a keystone species. We designed our program to monitor the density of beaver food caches (abundance) within sampling units with beaver and the proportion of sampling units with beavers present at the scale of a national forest. We designated watersheds as sampling units in a stratified random sampling design that we developed based on habitat modeling results. Habitat modeling indicated that the most suitable beaver habitat was near perennial water, near aspen (Populus tremuloides) and willow (Salix spp.), and in low gradient streams at lower elevations. Results from the initial monitoring period in October 2007 allowed us to assess costs and logistical considerations, validate our habitat model, and conduct power analyses to assess whether our sampling design could detect the level of declines in beaver stated in the monitoring objectives. Beaver food caches were located in 20 of 52 sampled watersheds. Monitoring 20 to 25 watersheds with beaver should provide sufficient power to detect 15-40% declines in the beaver food cache index as well as a twofold decline in the odds of beaver being present in watersheds. Indices of abundance, such as the beaver food cache index, provide a practical measure of

  2. End-to-End Study of the Transfer of Energy from Magnetosheath Ion Precipitation to the Cusp

    NASA Technical Reports Server (NTRS)

    Coffey, V. N.; Chandler, M. O.; Singh, Nagendra; Avanov, Levon

    2005-01-01

    This paper describes a study of the effects of unstable magnetosheath distributions on the cusp ionosphere. An end-to-end numerical model was used to study, first, the evolved distributions from precipitation due to reconnection and, secondly, the energy transfer into the high latitude ionosphere based on these solar wind/magnetosheath inputs. Using inputs of several representative examples of magnetosheath injections, waves were generated at the lower hybrid frequency and energy transferred to the ionospheric electrons and ions. The resulting wave spectra and ion and electron particle heating was analyzed. Keywords: Ion heating: Magnetosheath/Ionosphere coupling: Particle/Wave Interactions. Simulations

  3. An end-to-end system to identify temporal relation in discharge summaries: 2012 i2b2 challenge

    PubMed Central

    Xu, Yan; Wang, Yining; Liu, Tianren; Tsujii, Junichi; Chang, Eric I-Chao

    2013-01-01

    Objective To create an end-to-end system to identify temporal relation in discharge summaries for the 2012 i2b2 challenge. The challenge includes event extraction, timex extraction, and temporal relation identification. Design An end-to-end temporal relation system was developed. It includes three subsystems: an event extraction system (conditional random fields (CRF) name entity extraction and their corresponding attribute classifiers), a temporal extraction system (CRF name entity extraction, their corresponding attribute classifiers, and context-free grammar based normalization system), and a temporal relation system (10 multi-support vector machine (SVM) classifiers and a Markov logic networks inference system) using labeled sequential pattern mining, syntactic structures based on parse trees, and results from a coordination classifier. Micro-averaged precision (P), recall (R), averaged P&R (P&R), and F measure (F) were used to evaluate results. Results For event extraction, the system achieved 0.9415 (P), 0.8930 (R), 0.9166 (P&R), and 0.9166 (F). The accuracies of their type, polarity, and modality were 0.8574, 0.8585, and 0.8560, respectively. For timex extraction, the system achieved 0.8818, 0.9489, 0.9141, and 0.9141, respectively. The accuracies of their type, value, and modifier were 0.8929, 0.7170, and 0.8907, respectively. For temporal relation, the system achieved 0.6589, 0.7129, 0.6767, and 0.6849, respectively. For end-to-end temporal relation, it achieved 0.5904, 0.5944, 0.5921, and 0.5924, respectively. With the F measure used for evaluation, we were ranked first out of 14 competing teams (event extraction), first out of 14 teams (timex extraction), third out of 12 teams (temporal relation), and second out of seven teams (end-to-end temporal relation). Conclusions The system achieved encouraging results, demonstrating the feasibility of the tasks defined by the i2b2 organizers. The experiment result demonstrates that both global and local

  4. The Kepler End-to-End Model: Creating High-Fidelity Simulations to Test Kepler Ground Processing

    NASA Technical Reports Server (NTRS)

    Bryson, Stephen T.; Jenkins, Jon M.; Peters, Dan J.; Tenenbaum, Peter P.; Klaus, Todd C.; Gunter, Jay P.; Cote, Miles T.; Caldwell, Douglas A.

    2010-01-01

    The Kepler mission is designed to detect the transit of Earth-like planets around Sun-like stars by observing 100,000 stellar targets. Developing and testing the Kepler ground-segment processing system, in particular the data analysis pipeline, requires high-fidelity simulated data. This simulated data is provided by the Kepler End-to-End Model (ETEM). ETEM simulates the astrophysics of planetary transits and other phenomena, properties of the Kepler spacecraft and the format of the downlinked data. Major challenges addressed by ETEM include the rapid production of large amounts of simulated data, extensibility and maintainability.

  5. End-to-End Study of the Transfer of Energy from Magnetosheath Ion Precipitation to the Cusp

    NASA Technical Reports Server (NTRS)

    Coffey, V. N.; Chandler, M. O.; Singh, Nagendra; Avanov, Levon

    2005-01-01

    This paper describes a study of the effects of unstable magnetosheath distributions on the cusp ionosphere. An end-to-end numerical model was used to study, first, the evolved distributions from precipitation due to reconnection and, secondly, the energy transfer into the high latitude ionosphere based on these solar wind/magnetosheath inputs. Using inputs of several representative examples of magnetosheath injections, waves were generated at the lower hybrid frequency and energy transferred to the ionospheric electrons and ions. The resulting wave spectra and ion and electron particle heating was analyzed. Keywords: Ion heating: Magnetosheath/Ionosphere coupling: Particle/Wave Interactions. Simulations

  6. JADS JT&E: Phase 3 and Phase 4 Verification and Validation Report for the End-to-End Test

    DTIC Science & Technology

    1999-03-01

    Plan for the ETE Test and the Phase 3 Verification and Validation Plan for the End-to-End Test 1.2 Verification and Validation Tasks The V&V tasks...performed on 23 Ferumary and 13 March 1999 during Phase 3 were conducted on the T3 aircraft parked on the ramp and are described in the Phase 3 Verification...properly was to perform an abbreviated version of the Phase 3 V&V. There were also two V&V tasks that were either not completed or were not resolved when

  7. Automated Design of Propellant-Optimal, End-to-End, Low-Thrust Trajectories for Trojan Asteroid Tours

    NASA Technical Reports Server (NTRS)

    Stuart, Jeffrey; Howell, Kathleen; Wilson, Roby

    2013-01-01

    The Sun-Jupiter Trojan asteroids are celestial bodies of great scientific interest as well as potential resources offering water and other mineral resources for longterm human exploration of the solar system. Previous investigations under this project have addressed the automated design of tours within the asteroid swarm. This investigation expands the current automation scheme by incorporating options for a complete trajectory design approach to the Trojan asteroids. Computational aspects of the design procedure are automated such that end-to-end trajectories are generated with a minimum of human interaction after key elements and constraints associated with a proposed mission concept are specified.

  8. End-to-end testing. [to verify electrical equipment failure due to carbon fibers released in aircraft-fuel fires

    NASA Technical Reports Server (NTRS)

    Pride, R. A.

    1979-01-01

    The principle objective of the kinds of demonstration tests that are discussed is to try to verify whether or not carbon fibers that are released by burning composite parts in an aircraft-fuel fires can produce failures in electrical equipment. A secondary objective discussed is to experimentally validate the analytical models for some of the key elements in the risk analysis. The approach to this demonstration testing is twofold: limited end-to-end test are to be conducted in a shock tube; and planning for some large outdoor burn tests is being done.

  9. Far-Infrared Therapy Promotes Nerve Repair following End-to-End Neurorrhaphy in Rat Models of Sciatic Nerve Injury

    PubMed Central

    Chen, Tai-Yuan; Yang, Yi-Chin; Sha, Ya-Na; Chou, Jiun-Rou

    2015-01-01

    This study employed a rat model of sciatic nerve injury to investigate the effects of postoperative low-power far-infrared (FIR) radiation therapy on nerve repair following end-to-end neurorrhaphy. The rat models were divided into the following 3 groups: (1) nerve injury without FIR biostimulation (NI/sham group); (2) nerve injury with FIR biostimulation (NI/FIR group); and (3) noninjured controls (normal group). Walking-track analysis results showed that the NI/FIR group exhibited significantly higher sciatic functional indices at 8 weeks after surgery (P < 0.05) compared with the NI/sham group. The decreased expression of CD4 and CD8 in the NI/FIR group indicated that FIR irradiation modulated the inflammatory process during recovery. Compared with the NI/sham group, the NI/FIR group exhibited a significant reduction in muscle atrophy (P < 0.05). Furthermore, histomorphometric assessment indicated that the nerves regenerated more rapidly in the NI/FIR group than in the NI/sham group; furthermore, the NI/FIR group regenerated neural tissue over a larger area, as well as nerve fibers of greater diameter and with thicker myelin sheaths. Functional recovery, inflammatory response, muscular reinnervation, and histomorphometric assessment all indicated that FIR radiation therapy can accelerate nerve repair following end-to-end neurorrhaphy of the sciatic nerve. PMID:25722734

  10. Effect of swirling flow on platelet concentration distribution in small-caliber artificial grafts and end-to-end anastomoses

    NASA Astrophysics Data System (ADS)

    Zhan, Fan; Fan, Yu-Bo; Deng, Xiao-Yan

    2011-10-01

    Platelet concentration near the blood vessel wall is one of the major factors in the adhesion of platelets to the wall. In our previous studies, it was found that swirling flows could suppress platelet adhesion in small-caliber artificial grafts and end-to-end anastomoses. In order to better understand the beneficial effect of the swirling flow, we numerically analyzed the near-wall concentration distribution of platelets in a straight tube and a sudden tubular expansion tube under both swirling flow and normal flow conditions. The numerical models were created based on our previous experimental studies. The simulation results revealed that when compared with the normal flow, the swirling flow could significantly reduce the near-wall concentration of platelets in both the straight tube and the expansion tube. The present numerical study therefore indicates that the reduction in platelet adhesion under swirling flow conditions in small-caliber arterial grafts, or in end-to-end anastomoses as observed in our previous experimental study, was possibly through a mechanism of platelet transport, in which the swirling flow reduced the near-wall concentration of platelets.

  11. End-to-end self-assembly of gold nanorods in isopropanol solution: experimental and theoretical studies

    NASA Astrophysics Data System (ADS)

    Gordel, M.; Piela, K.; Kołkowski, R.; Koźlecki, T.; Buckle, M.; Samoć, M.

    2015-12-01

    We describe here a modification of properties of colloidal gold nanorods (NRs) resulting from the chemical treatment used to carry out their transfer into isopropanol (IPA) solution. The NRs acquire a tendency to attach one to another by their ends (end-to-end assembly). We focus on the investigation of the change in position and shape of the longitudinal surface plasmon (l-SPR) band after self-assembly. The experimental results are supported by a theoretical calculation, which rationalizes the dramatic change in optical properties when the NRs are positioned end-to-end at short distances. The detailed spectroscopic characterization performed at the consecutive stages of transfer of the NRs from water into IPA solution revealed the features of the interaction between the polymers used as ligands and their contribution to the final stage, when the NRs were dispersed in IPA solution. The efficient method of aligning the NRs detailed here may facilitate applications of the self-assembled NRs as building blocks for optical materials and biological sensing.

  12. Image gathering, coding, and processing: End-to-end optimization for efficient and robust acquisition of visual information

    NASA Technical Reports Server (NTRS)

    Huck, Friedrich O.; Fales, Carl L.

    1990-01-01

    Researchers are concerned with the end-to-end performance of image gathering, coding, and processing. The applications range from high-resolution television to vision-based robotics, wherever the resolution, efficiency and robustness of visual information acquisition and processing are critical. For the presentation at this workshop, it is convenient to divide research activities into the following two overlapping areas: The first is the development of focal-plane processing techniques and technology to effectively combine image gathering with coding, with an emphasis on low-level vision processing akin to the retinal processing in human vision. The approach includes the familiar Laplacian pyramid, the new intensity-dependent spatial summation, and parallel sensing/processing networks. Three-dimensional image gathering is attained by combining laser ranging with sensor-array imaging. The second is the rigorous extension of information theory and optimal filtering to visual information acquisition and processing. The goal is to provide a comprehensive methodology for quantitatively assessing the end-to-end performance of image gathering, coding, and processing.

  13. End-To-End Risk Assesment: From Genes and Protein to Acceptable Radiation Risks for Mars Exploration

    NASA Technical Reports Server (NTRS)

    Cucinotta, Francis A.; Schimmerling, Walter

    2000-01-01

    The human exploration of Mars will impose unavoidable health risks from galactic cosmic rays (GCR) and possibly solar particle events (SPE). It is the goal of NASA's Space Radiation Health Program to develop the capability to predict health risks with significant accuracy to ensure that risks are well below acceptable levels and to allow for mitigation approaches to be effective at reasonable costs. End-to-End risk assessment is the approach being followed to understand proton and heavy ion damage at the molecular, cellular, and tissue levels in order to predict the probability of the major health risk including cancer, neurological disorders, hereditary effects, cataracts, and acute radiation sickness and to develop countermeasures for mitigating risks.

  14. End-to-End Trajectory for Conjunction Class Mars Missions Using Hybrid Solar-Electric/Chemical Transportation System

    NASA Technical Reports Server (NTRS)

    Chai, Patrick R.; Merrill, Raymond G.; Qu, Min

    2016-01-01

    NASA's Human Spaceflight Architecture Team is developing a reusable hybrid transportation architecture in which both chemical and solar-electric propulsion systems are used to deliver crew and cargo to exploration destinations. By combining chemical and solar-electric propulsion into a single spacecraft and applying each where it is most effective, the hybrid architecture enables a series of Mars trajectories that are more fuel efficient than an all chemical propulsion architecture without significant increases to trip time. The architecture calls for the aggregation of exploration assets in cislunar space prior to departure for Mars and utilizes high energy lunar-distant high Earth orbits for the final staging prior to departure. This paper presents the detailed analysis of various cislunar operations for the EMC Hybrid architecture as well as the result of the higher fidelity end-to-end trajectory analysis to understand the implications of the design choices on the Mars exploration campaign.

  15. End-To-End Risk Assesment: From Genes and Protein to Acceptable Radiation Risks for Mars Exploration

    NASA Technical Reports Server (NTRS)

    Cucinotta, Francis A.; Schimmerling, Walter

    2000-01-01

    The human exploration of Mars will impose unavoidable health risks from galactic cosmic rays (GCR) and possibly solar particle events (SPE). It is the goal of NASA's Space Radiation Health Program to develop the capability to predict health risks with significant accuracy to ensure that risks are well below acceptable levels and to allow for mitigation approaches to be effective at reasonable costs. End-to-End risk assessment is the approach being followed to understand proton and heavy ion damage at the molecular, cellular, and tissue levels in order to predict the probability of the major health risk including cancer, neurological disorders, hereditary effects, cataracts, and acute radiation sickness and to develop countermeasures for mitigating risks.

  16. HITSZ_CDR: an end-to-end chemical and disease relation extraction system for BioCreative V

    PubMed Central

    Li, Haodi; Tang, Buzhou; Chen, Qingcai; Chen, Kai; Wang, Xiaolong; Wang, Baohua; Wang, Zhe

    2016-01-01

    In this article, an end-to-end system was proposed for the challenge task of disease named entity recognition (DNER) and chemical-induced disease (CID) relation extraction in BioCreative V, where DNER includes disease mention recognition (DMR) and normalization (DN). Evaluation on the challenge corpus showed that our system achieved the highest F1-scores 86.93% on DMR, 84.11% on DN, 43.04% on CID relation extraction, respectively. The F1-score on DMR is higher than our previous one reported by the challenge organizers (86.76%), the highest F1-score of the challenge. Database URL: http://database.oxfordjournals.org/content/2016/baw077 PMID:27270713

  17. The MARS pathfinder end-to-end information system: A pathfinder for the development of future NASA planetary missions

    NASA Technical Reports Server (NTRS)

    Cook, Richard A.; Kazz, Greg J.; Tai, Wallace S.

    1996-01-01

    The development of the Mars pathfinder is considered with emphasis on the End-to-End Information System (EEIS) development approach. The primary mission objective is to successfully develop and deliver a single flight system to the Martian surface, demonstrating entry, descent and landing. The EEIS is a set of functions distributed throughout the flight, ground and Mission Operation Systems (MOS) that inter-operate in order to control, collect, transport, process, store and analyze the uplink and downlink information flows of the mission. Coherence between the mission systems is achieved though the EEIS architecture. The key characteristics of the system are: a concurrent engineering approach for the development of flight, ground and mission operation systems; the fundamental EEIS architectural heuristics; a phased incremental EEIS development and test approach, and an EEIS design deploying flight, ground and MOS operability features, including integrated ground and flight based toolsets.

  18. The MARS pathfinder end-to-end information system: A pathfinder for the development of future NASA planetary missions

    NASA Technical Reports Server (NTRS)

    Cook, Richard A.; Kazz, Greg J.; Tai, Wallace S.

    1996-01-01

    The development of the Mars pathfinder is considered with emphasis on the End-to-End Information System (EEIS) development approach. The primary mission objective is to successfully develop and deliver a single flight system to the Martian surface, demonstrating entry, descent and landing. The EEIS is a set of functions distributed throughout the flight, ground and Mission Operation Systems (MOS) that inter-operate in order to control, collect, transport, process, store and analyze the uplink and downlink information flows of the mission. Coherence between the mission systems is achieved though the EEIS architecture. The key characteristics of the system are: a concurrent engineering approach for the development of flight, ground and mission operation systems; the fundamental EEIS architectural heuristics; a phased incremental EEIS development and test approach, and an EEIS design deploying flight, ground and MOS operability features, including integrated ground and flight based toolsets.

  19. NASA End-to-End Data System /NEEDS/ information adaptive system - Performing image processing onboard the spacecraft

    NASA Technical Reports Server (NTRS)

    Kelly, W. L.; Howle, W. M.; Meredith, B. D.

    1980-01-01

    The Information Adaptive System (IAS) is an element of the NASA End-to-End Data System (NEEDS) Phase II and is focused toward onbaord image processing. Since the IAS is a data preprocessing system which is closely coupled to the sensor system, it serves as a first step in providing a 'Smart' imaging sensor. Some of the functions planned for the IAS include sensor response nonuniformity correction, geometric correction, data set selection, data formatting, packetization, and adaptive system control. The inclusion of these sensor data preprocessing functions onboard the spacecraft will significantly improve the extraction of information from the sensor data in a timely and cost effective manner and provide the opportunity to design sensor systems which can be reconfigured in near real time for optimum performance. The purpose of this paper is to present the preliminary design of the IAS and the plans for its development.

  20. The initial data products from the EUVE software - A photon's journey through the End-to-End System

    NASA Technical Reports Server (NTRS)

    Antia, Behram

    1993-01-01

    The End-to-End System (EES) is a unique collection of software modules created for use at the Center for EUV Astrophysics. The 'pipeline' is a shell script which executes selected EES modules and creates initial data products: skymaps, data sets for individual sources (called 'pigeonholes') and catalogs of sources. This article emphasizes the data from the all-sky survey, conducted between July 22, 1992 and January 21, 1993. A description of each of the major data products will be given and, as an example of how the pipeline works, the reader will follow a photon's path through the software pipeline into a pigeonhole. These data products are the primary goal of the EUVE all-sky survey mission, and so their relative importance for the follow-up science will also be discussed.

  1. SU-E-T-282: Dose Measurements with An End-To-End Audit Phantom for Stereotactic Radiotherapy

    SciTech Connect

    Jones, R; Artschan, R; Thwaites, D; Lehmann, J

    2015-06-15

    Purpose: Report on dose measurements as part of an end-to-end test for stereotactic radiotherapy, using a new audit tool, which allows audits to be performed efficiently either by an onsite team or as a postal audit. Methods: Film measurements have been performed with a new Stereotactic Cube Phantom. The phantom has been designed to perform Winston Lutz type position verification measurements and dose measurements in one setup. It comprises a plastic cube with a high density ball in its centre (used for MV imaging with film or EPID) and low density markers in the periphery (used for Cone Beam Computed Tomography, CBCT imaging). It also features strategically placed gold markers near the posterior and right surfaces, which can be used to calculate phantom rotations on MV images. Slit-like openings allow insertion of film or other detectors.The phantom was scanned and small field treatment plans were created. The fields do not traverse any inhomogeneities of the phantom on their paths to the measurement location. The phantom was setup at the delivery system using CBCT imaging. The calculated treatment fields were delivered, each with a piece of radiochromic film (EBT3) placed in the anterior film holder of the phantom. MU had been selected in planning to achieve similar exposures on all films. Calibration films were exposed in solid water for dose levels around the expected doses. Films were scanned and analysed following established procedures. Results: Setup of the cube showed excellent suitability for CBCT 3D alignment. MV imaging with EPID allowed for clear identification of all markers. Film based dose measurements showed good agreement for MLC created fields down to 0.5 mm × 0.5 mm. Conclusion: An end-to-end audit phantom for stereotactic radiotherapy has been developed and tested.

  2. Distributed Large Data-Object Environments: End-to-End Performance Analysis of High Speed Distributed Storage Systems in Wide Area ATM Networks

    NASA Technical Reports Server (NTRS)

    Johnston, William; Tierney, Brian; Lee, Jason; Hoo, Gary; Thompson, Mary

    1996-01-01

    We have developed and deployed a distributed-parallel storage system (DPSS) in several high speed asynchronous transfer mode (ATM) wide area networks (WAN) testbeds to support several different types of data-intensive applications. Architecturally, the DPSS is a network striped disk array, but is fairly unique in that its implementation allows applications complete freedom to determine optimal data layout, replication and/or coding redundancy strategy, security policy, and dynamic reconfiguration. In conjunction with the DPSS, we have developed a 'top-to-bottom, end-to-end' performance monitoring and analysis methodology that has allowed us to characterize all aspects of the DPSS operating in high speed ATM networks. In particular, we have run a variety of performance monitoring experiments involving the DPSS in the MAGIC testbed, which is a large scale, high speed, ATM network and we describe our experience using the monitoring methodology to identify and correct problems that limit the performance of high speed distributed applications. Finally, the DPSS is part of an overall architecture for using high speed, WAN's for enabling the routine, location independent use of large data-objects. Since this is part of the motivation for a distributed storage system, we describe this architecture.

  3. End-to-End Information System design at the NASA Jet Propulsion Laboratory. [data transmission between user and space-based sensor

    NASA Technical Reports Server (NTRS)

    Hooke, A. J.

    1978-01-01

    In recognition of a pressing need of the 1980s to optimize the two-way flow of information between a ground-based user and a remote-space-based sensor, an end-to-end approach to the design of information systems has been adopted at the JPL. This paper reviews End-to-End Information System (EEIS) activity at the JPL, with attention given to the scope of the EEIS transfer function, and functional and physical elements of the EEIS. The relationship between the EEIS and the NASA End-to-End Data System program is discussed.

  4. End-to-End Information System design at the NASA Jet Propulsion Laboratory. [data transmission between user and space-based sensor

    NASA Technical Reports Server (NTRS)

    Hooke, A. J.

    1978-01-01

    In recognition of a pressing need of the 1980s to optimize the two-way flow of information between a ground-based user and a remote-space-based sensor, an end-to-end approach to the design of information systems has been adopted at the JPL. This paper reviews End-to-End Information System (EEIS) activity at the JPL, with attention given to the scope of the EEIS transfer function, and functional and physical elements of the EEIS. The relationship between the EEIS and the NASA End-to-End Data System program is discussed.

  5. Thermodynamic state retrieval from microwave occultation data and performance analysis based on end-to-end simulations

    NASA Astrophysics Data System (ADS)

    Schweitzer, S.; Kirchengast, G.; Schwaerz, M.; Fritzer, J.; Gorbunov, M. E.

    2011-05-01

    Microwave occultation using centimeter- and millimeter-wave signals between low Earth orbit (LEO) satellites (LEO microwave occultation, LMO) is an advancement of GPS radio occultation (GRO) exploiting in addition to refraction also absorption of signals. Beyond the successful GRO refractivity profiling capability, which leaves a temperature-humidity ambiguity in the troposphere where moisture cannot be neglected, LMO enables joint retrieval of pressure, temperature, and humidity profiles without auxiliary background information. Here we focus on the upper troposphere/lower stratosphere and advance the LMO method in two ways: (1) we introduce a new retrieval algorithm for processing LMO excess phase and amplitude data from multiple frequencies, complementing existing GRO retrieval algorithms, and (2) we employ the algorithm in an ensemble-based end-to-end performance analysis and assess the accuracy of pressure, temperature, and humidity profiles retrieved from the LMO data. The end-to-end simulations were carried out under quasi-realistic conditions for a day of LEO-LEO occultation events, based on a high-resolution atmospheric analysis of the European Centre for Medium-Range Weather Forecasts (ECMWF) and accounting for scintillation noise from turbulence and instrumental errors. The new algorithm was found robust, fast, and versatile to adequately process LMO data under all conditions from dry and clear to moist and cloudy air as contained in the ECMWF analysis. The retrieved pressure, temperature, and humidity profiles were generally found unbiased and within target accuracy requirements, set by scientific objectives of atmosphere and climate research going to be supported by the data, of <0.2% (pressure), <0.5 K (temperature), and <10% (humidity). Extending a "minimum" LMO design with three frequencies near 22 GHz with two added frequencies near 183 GHz favorably provides humidity retrieval into the lower stratosphere but already the "minimum" design resolves

  6. A novel end-to-end classifier using domain transferred deep convolutional neural networks for biomedical images.

    PubMed

    Pang, Shuchao; Yu, Zhezhou; Orgun, Mehmet A

    2017-03-01

    Highly accurate classification of biomedical images is an essential task in the clinical diagnosis of numerous medical diseases identified from those images. Traditional image classification methods combined with hand-crafted image feature descriptors and various classifiers are not able to effectively improve the accuracy rate and meet the high requirements of classification of biomedical images. The same also holds true for artificial neural network models directly trained with limited biomedical images used as training data or directly used as a black box to extract the deep features based on another distant dataset. In this study, we propose a highly reliable and accurate end-to-end classifier for all kinds of biomedical images via deep learning and transfer learning. We first apply domain transferred deep convolutional neural network for building a deep model; and then develop an overall deep learning architecture based on the raw pixels of original biomedical images using supervised training. In our model, we do not need the manual design of the feature space, seek an effective feature vector classifier or segment specific detection object and image patches, which are the main technological difficulties in the adoption of traditional image classification methods. Moreover, we do not need to be concerned with whether there are large training sets of annotated biomedical images, affordable parallel computing resources featuring GPUs or long times to wait for training a perfect deep model, which are the main problems to train deep neural networks for biomedical image classification as observed in recent works. With the utilization of a simple data augmentation method and fast convergence speed, our algorithm can achieve the best accuracy rate and outstanding classification ability for biomedical images. We have evaluated our classifier on several well-known public biomedical datasets and compared it with several state-of-the-art approaches. We propose a robust

  7. End-to-end Cyberinfrastructure and Data Services for Earth System Science Education and Research: A vision for the future

    NASA Astrophysics Data System (ADS)

    Ramamurthy, M. K.

    2006-05-01

    yet revolutionary way of building applications and methods to connect and exchange information over the Web. This new approach, based on XML - a widely accepted format for exchanging data and corresponding semantics over the Internet - enables applications, computer systems, and information processes to work together in fundamentally different ways. Likewise, the advent of digital libraries, grid computing platforms, interoperable frameworks, standards and protocols, open-source software, and community atmospheric models have been important drivers in shaping the use of a new generation of end-to-end cyberinfrastructure for solving some of the most challenging scientific and educational problems. In this talk, I will present an overview of the scientific, technological, and educational landscape, discuss recent developments in cyberinfrastructure, and Unidata's role in and vision for providing easy-to use, robust, end-to-end data services for solving geoscientific problems and advancing student learning.

  8. End-to-end Cyberinfrastructure and Data Services for Earth System Science Education and Research: Unidata's Plans and Directions

    NASA Astrophysics Data System (ADS)

    Ramamurthy, M.

    2005-12-01

    work together in a fundamentally different way. Likewise, the advent of digital libraries, grid computing platforms, interoperable frameworks, standards and protocols, open-source software, and community atmospheric models have been important drivers in shaping the use of a new generation of end-to-end cyberinfrastructure for solving some of the most challenging scientific and educational problems. In this talk, I will present an overview of the scientific, technological, and educational drivers and discuss recent developments in cyberinfrastructure and Unidata's role and directions in providing robust, end-to-end data services for solving geoscientific problems and advancing student learning.

  9. Towards end-to-end models for investigating the effects of climate and fishing in marine ecosystems

    NASA Astrophysics Data System (ADS)

    Travers, M.; Shin, Y.-J.; Jennings, S.; Cury, P.

    2007-12-01

    End-to-end models that represent ecosystem components from primary producers to top predators, linked through trophic interactions and affected by the abiotic environment, are expected to provide valuable tools for assessing the effects of climate change and fishing on ecosystem dynamics. Here, we review the main process-based approaches used for marine ecosystem modelling, focusing on the extent of the food web modelled, the forcing factors considered, the trophic processes represented, as well as the potential use and further development of the models. We consider models of a subset of the food web, models which represent the first attempts to couple low and high trophic levels, integrated models of the whole ecosystem, and size spectrum models. Comparisons within and among these groups of models highlight the preferential use of functional groups at low trophic levels and species at higher trophic levels and the different ways in which the models account for abiotic processes. The model comparisons also highlight the importance of choosing an appropriate spatial dimension for representing organism dynamics. Many of the reviewed models could be extended by adding components and by ensuring that the full life cycles of species components are represented, but end-to-end models should provide full coverage of ecosystem components, the integration of physical and biological processes at different scales and two-way interactions between ecosystem components. We suggest that this is best achieved by coupling models, but there are very few existing cases where the coupling supports true two-way interaction. The advantages of coupling models are that the extent of discretization and representation can be targeted to the part of the food web being considered, making their development time- and cost-effective. Processes such as predation can be coupled to allow the propagation of forcing factors effects up and down the food web. However, there needs to be a stronger focus

  10. Unmanned Aircraft Systems Minimum Operations Performance Standards End-to-End Verification and Validation (E2-V2) Simulation

    NASA Technical Reports Server (NTRS)

    Ghatas, Rania W.; Jack, Devin P.; Tsakpinis, Dimitrios; Vincent, Michael J.; Sturdy, James L.; Munoz, Cesar A.; Hoffler, Keith D.; Dutle, Aaron M.; Myer, Robert R.; Dehaven, Anna M.; hide

    2017-01-01

    As Unmanned Aircraft Systems (UAS) make their way to mainstream aviation operations within the National Airspace System (NAS), research efforts are underway to develop a safe and effective environment for their integration into the NAS. Detect and Avoid (DAA) systems are required to account for the lack of "eyes in the sky" due to having no human on-board the aircraft. The current NAS relies on pilot's vigilance and judgement to remain Well Clear (CFR 14 91.113) of other aircraft. RTCA SC-228 has defined DAA Well Clear (DAAWC) to provide a quantified Well Clear volume to allow systems to be designed and measured against. Extended research efforts have been conducted to understand and quantify system requirements needed to support a UAS pilot's ability to remain well clear of other aircraft. The efforts have included developing and testing sensor, algorithm, alerting, and display requirements. More recently, sensor uncertainty and uncertainty mitigation strategies have been evaluated. This paper discusses results and lessons learned from an End-to-End Verification and Validation (E2-V2) simulation study of a DAA system representative of RTCA SC-228's proposed Phase I DAA Minimum Operational Performance Standards (MOPS). NASA Langley Research Center (LaRC) was called upon to develop a system that evaluates a specific set of encounters, in a variety of geometries, with end-to-end DAA functionality including the use of sensor and tracker models, a sensor uncertainty mitigation model, DAA algorithmic guidance in both vertical and horizontal maneuvering, and a pilot model which maneuvers the ownship aircraft to remain well clear from intruder aircraft, having received collective input from the previous modules of the system. LaRC developed a functioning batch simulation and added a sensor/tracker model from the Federal Aviation Administration (FAA) William J. Hughes Technical Center, an in-house developed sensor uncertainty mitigation strategy, and implemented a pilot

  11. A novel end-to-end fault detection and localization protocol for wavelength-routed WDM networks

    NASA Astrophysics Data System (ADS)

    Zeng, Hongqing; Vukovic, Alex; Huang, Changcheng

    2005-09-01

    Recently the wavelength division multiplexing (WDM) networks are becoming prevalent for telecommunication networks. However, even a very short disruption of service caused by network faults may lead to high data loss in such networks due to the high date rates, increased wavelength numbers and density. Therefore, the network survivability is critical and has been intensively studied, where fault detection and localization is the vital part but has received disproportional attentions. In this paper we describe and analyze an end-to-end lightpath fault detection scheme in data plane with the fault notification in control plane. The endeavor is focused on reducing the fault detection time. In this protocol, the source node of each lightpath keeps sending hello packets to the destination node exactly following the path for data traffic. The destination node generates an alarm once a certain number of consecutive hello packets are missed within a given time period. Then the network management unit collects all alarms and locates the faulty source based on the network topology, as well as sends fault notification messages via control plane to either the source node or all upstream nodes along the lightpath. The performance evaluation shows such a protocol can achieve fast fault detection, and at the same time, the overhead brought to the user data by hello packets is negligible.

  12. End-to-end simulation of high-contrast imaging systems: methods and results for the PICTURE mission family

    NASA Astrophysics Data System (ADS)

    Douglas, Ewan S.; Hewasawam, Kuravi; Mendillo, Christopher B.; Cahoy, Kerri L.; Cook, Timothy A.; Finn, Susanna C.; Howe, Glenn A.; Kuchner, Marc J.; Lewis, Nikole K.; Marinan, Anne D.; Mawet, Dimitri; Chakrabarti, Supriya

    2015-09-01

    We describe a set of numerical approaches to modeling the performance of space flight high-contrast imaging payloads. Mission design for high-contrast imaging requires numerical wavefront error propagation to ensure accurate component specifications. For constructed instruments, wavelength and angle-dependent throughput and contrast models allow detailed simulations of science observations, allowing mission planners to select the most productive science targets. The PICTURE family of missions seek to quantify the optical brightness of scattered light from extrasolar debris disks via several high-contrast imaging techniques: sounding rocket (the Planet Imaging Concept Testbed Using a Rocket Experiment) and balloon flights of a visible nulling coronagraph, as well as a balloon flight of a vector vortex coronagraph (the Planetary Imaging Concept Testbed Using a Recoverable Experiment - Coronagraph, PICTURE-C). The rocket mission employs an on-axis 0.5m Gregorian telescope, while the balloon flights will share an unobstructed off-axis 0.6m Gregorian. This work details the flexible approach to polychromatic, end-to-end physical optics simulations used for both the balloon vector vortex coronagraph and rocket visible nulling coronagraph missions. We show the preliminary PICTURE-C telescope and vector vortex coronagraph design will achieve 10-8 contrast without post-processing as limited by realistic optics, but not considering polarization or low-order errors. Simulated science observations of the predicted warm ring around Epsilon Eridani illustrate the performance of both missions.

  13. Saphenous vein to internal mammary artery end-to-end composite grafts for coronary artery bypass. Late follow-up.

    PubMed

    Paterson, Hugh S; Thakkar, Jay; Byth, Karen; Denniss, A Robert

    2015-02-01

    Internal mammary artery (IMA) grafts provide equal or superior graft patency compared to other conduits. The IMA length limits extensive myocardial revascularisation with IMA grafts alone. This study aimed to determine the results of lengthening free IMAs with a short proximal segment of saphenous vein (SV) to enable more extensive myocardial revascularisation. Patients (n=92) who underwent end-to-end composite SV-IMA grafts were followed up through cardiology and death register databases. The mean patient age was 57.5 years and median follow up 10.9 years. There was no perioperative mortality and 10-year survival was 89.6%. Thirty-one patients (34%) underwent repeat angiography at a median of 2.8 years postoperatively. The 10-year freedom from angiography showing SV segment occlusion was 89% with a median time to angiography of 2.3 years (nine patients). The number of distal anastomoses was the only independent predictor of SV segment occlusion HR per anastomosis=0.26 (p=0.01). In five sequential grafts to the circumflex and right coronary systems, the IMA portion of the graft remained patent following SV segment occlusion. Graft patency is improved by a greater number of coronary artery anastomoses. Copyright © 2014 Australian and New Zealand Society of Cardiac and Thoracic Surgeons (ANZSCTS) and the Cardiac Society of Australia and New Zealand (CSANZ). Published by Elsevier B.V. All rights reserved.

  14. Hardware and Methods of the Optical End-to-End Test of the Far Ultraviolet Spectroscopic Explorer (FUSE)

    NASA Technical Reports Server (NTRS)

    Conard, Steven J.; Redman, Kevin W.; Barkhouser, Robert H.; McGuffey, Doug B.; Smee, Stephen; Ohl, Raymond G.; Kushner, Gary

    1999-01-01

    The Far Ultraviolet Spectroscopic Explorer (FUSE), currently being tested and scheduled for a 1999 launch, is an astrophysics satellite designed to provide high spectral resolving power (Lambda/(Delta)Lambda = 24,000-30,000) over the interval 90.5-118.7 nm. The FUSE optical path consists of four co-aligned, normal incidence, off-axis parabolic, primary mirrors which illuminate separate Rowland circle spectrograph channels equipped with holographic gratings and delay line microchannel plate detectors. We describe the hardware and methods used for the optical end-to-end test of the FUSE instrument during satellite integration and test. Cost and schedule constraints forced us to devise a simplified version of the planned optical test which occurred in parallel with satellite thermal-vacuum testing. The optical test employed a collimator assembly which consisted of four co-aligned, 15" Cassegrain telescopes which were positioned above the FUSE instrument, providing a collimated beam for each optical channel. A windowed UV light source, remotely adjustable in three axes, was mounted at the focal plane of each collimator. Problems with the UV light sources, including high F-number and window failures, were the only major difficulties encountered during the test. The test succeeded in uncovering a significant problem with the secondary structure used for the instrument closeout cavity and, furthermore, showed that the mechanical solution was successful. The hardware was also used extensively for simulations of science observations, providing both UV light for spectra and visible light for the fine error sensor camera.

  15. Scaffold-integrated microchips for end-to-end in vitro tumor cell attachment and xenograft formation

    PubMed Central

    Lee, Jungwoo; Kohl, Nathaniel; Shanbhang, Sachin; Parekkadan, Biju

    2015-01-01

    Microfluidic technologies have substantially advanced cancer research by enabling the isolation of rare circulating tumor cells (CTCs) for diagnostic and prognostic purposes. The characterization of isolated CTCs has been limited due to the difficulty in recovering and growing isolated cells with high fidelity. Here, we present a strategy that uses a 3D scaffold, integrated into a microfludic device, as a transferable substrate that can be readily isolated after device operation for serial use in vivo as a transplanted tissue bed. Hydrogel scaffolds were incorporated into a PDMS fluidic chamber prior to bonding and were rehydrated in the chamber after fluid contact. The hydrogel matrix completely filled the fluid chamber, significantly increasing the surface area to volume ratio, and could be directly visualized under a microscope. Computational modeling defined different flow and pressure regimes that guided the conditions used to operate the chip. As a proof of concept using a model cell line, we confirmed human prostate tumor cell attachment in the microfluidic scaffold chip, retrieval of the scaffold en masse, and serial implantation of the scaffold to a mouse model with preserved xenograft development. With further improvement in capture efficiency, this approach can offer an end-to-end platform for the continuous study of isolated cancer cells from a biological fluid to a xenograft in mice. PMID:26709385

  16. Integrated end-to-end metrology and data analysis system for the advanced x-ray astrophysics facility telescope mirrors

    NASA Astrophysics Data System (ADS)

    Sarnik, Andrea; Zimmerman, Gerry

    1992-01-01

    The Advanced X-ray Astrophysics Facility (AXAF) consists of a nested set of six Wolter Type 1 x-ray telescopes. Each telescope consists of two mirrors (a parabola and a hyperbola). The high resolution optical performance required by AXAF and the size of the mirrors necessitates enormous quantities of data to characterize the optics. We will describe an end-to-end data system to be used for the metrology and fabrication of these 12 mirrors. The data system must have the capability of collecting optic metrology data from several instruments, processing and analyzing data, and generating machine instructions for the next grinding or polishing cycle. This system consists of personal computers interfaced to metrology instruments for the automatic collection of data, personal computers that control grinder/polishers, a mainframe computer for storing and managing data, and workstations for data processing and analysis. All of these computers are networked together to facilitate data transfer between computers. The system also includes an extensive library of software whose functions include processing mechanical and interferometric measurements, fitting polynomials to the data, performing frequency analysis of the data, and doing performance predictions. This data system has been used in the fabrication of the first two AXAF mirrors, produced by Hughes Danbury under contract to TRW. These mirrors are the first in a telescope that will be well beyond the performance of any existing x-ray telescopes.

  17. An end-to-end software solution for the analysis of high-throughput single-cell migration data

    PubMed Central

    Masuzzo, Paola; Huyck, Lynn; Simiczyjew, Aleksandra; Ampe, Christophe; Martens, Lennart; Van Troys, Marleen

    2017-01-01

    The systematic study of single-cell migration requires the availability of software for assisting data inspection, quality control and analysis. This is especially important for high-throughput experiments, where multiple biological conditions are tested in parallel. Although the field of cell migration can count on different computational tools for cell segmentation and tracking, downstream data visualization, parameter extraction and statistical analysis are still left to the user and are currently not possible within a single tool. This article presents a completely new module for the open-source, cross-platform CellMissy software for cell migration data management. This module is the first tool to focus specifically on single-cell migration data downstream of image processing. It allows fast comparison across all tested conditions, providing automated data visualization, assisted data filtering and quality control, extraction of various commonly used cell migration parameters, and non-parametric statistical analysis. Importantly, the module enables parameters computation both at the trajectory- and at the step-level. Moreover, this single-cell analysis module is complemented by a new data import module that accommodates multiwell plate data obtained from high-throughput experiments, and is easily extensible through a plugin architecture. In conclusion, the end-to-end software solution presented here tackles a key bioinformatics challenge in the cell migration field, assisting researchers in their high-throughput data processing. PMID:28205527

  18. WARP (workflow for automated and rapid production): a framework for end-to-end automated digital print workflows

    NASA Astrophysics Data System (ADS)

    Joshi, Parag

    2006-02-01

    Publishing industry is experiencing a major paradigm shift with the advent of digital publishing technologies. A large number of components in the publishing and print production workflow are transformed in this shift. However, the process as a whole requires a great deal of human intervention for decision making and for resolving exceptions during job execution. Furthermore, a majority of the best-of-breed applications for publishing and print production are intrinsically designed and developed to be driven by humans. Thus, the human-intensive nature of the current prepress process accounts for a very significant amount of the overhead costs in fulfillment of jobs on press. It is a challenge to automate the functionality of applications built with the model of human driven exectution. Another challenge is to orchestrate various components in the publishing and print production pipeline such that they work in a seamless manner to enable the system to perform automatic detection of potential failures and take corrective actions in a proactive manner. Thus, there is a great need for a coherent and unifying workflow architecture that streamlines the process and automates it as a whole in order to create an end-to-end digital automated print production workflow that does not involve any human intervention. This paper describes an architecture and building blocks that lay the foundation for a plurality of automated print production workflows.

  19. A real-time 3D end-to-end augmented reality system (and its representation transformations)

    NASA Astrophysics Data System (ADS)

    Tytgat, Donny; Aerts, Maarten; De Busser, Jeroen; Lievens, Sammy; Rondao Alface, Patrice; Macq, Jean-Francois

    2016-09-01

    The new generation of HMDs coming to the market is expected to enable many new applications that allow free viewpoint experiences with captured video objects. Current applications usually rely on 3D content that is manually created or captured in an offline manner. In contrast, this paper focuses on augmented reality applications that use live captured 3D objects while maintaining free viewpoint interaction. We present a system that allows live dynamic 3D objects (e.g. a person who is talking) to be captured in real-time. Real-time performance is achieved by traversing a number of representation formats and exploiting their specific benefits. For instance, depth images are maintained for fast neighborhood retrieval and occlusion determination, while implicit surfaces are used to facilitate multi-source aggregation for both geometry and texture. The result is a 3D reconstruction system that outputs multi-textured triangle meshes at real-time rates. An end-to-end system is presented that captures and reconstructs live 3D data and allows for this data to be used on a networked (AR) device. For allocating the different functional blocks onto the available physical devices, a number of alternatives are proposed considering the available computational power and bandwidth for each of the components. As we will show, the representation format can play an important role in this functional allocation and allows for a flexible system that can support a highly heterogeneous infrastructure.

  20. Scaffold-integrated microchips for end-to-end in vitro tumor cell attachment and xenograft formation.

    PubMed

    Lee, Jungwoo; Kohl, Nathaniel; Shanbhang, Sachin; Parekkadan, Biju

    2015-12-01

    Microfluidic technologies have substantially advanced cancer research by enabling the isolation of rare circulating tumor cells (CTCs) for diagnostic and prognostic purposes. The characterization of isolated CTCs has been limited due to the difficulty in recovering and growing isolated cells with high fidelity. Here, we present a strategy that uses a 3D scaffold, integrated into a microfludic device, as a transferable substrate that can be readily isolated after device operation for serial use in vivo as a transplanted tissue bed. Hydrogel scaffolds were incorporated into a PDMS fluidic chamber prior to bonding and were rehydrated in the chamber after fluid contact. The hydrogel matrix completely filled the fluid chamber, significantly increasing the surface area to volume ratio, and could be directly visualized under a microscope. Computational modeling defined different flow and pressure regimes that guided the conditions used to operate the chip. As a proof of concept using a model cell line, we confirmed human prostate tumor cell attachment in the microfluidic scaffold chip, retrieval of the scaffold en masse, and serial implantation of the scaffold to a mouse model with preserved xenograft development. With further improvement in capture efficiency, this approach can offer an end-to-end platform for the continuous study of isolated cancer cells from a biological fluid to a xenograft in mice.

  1. Minimizing Barriers in Learning for On-Call Radiology Residents-End-to-End Web-Based Resident Feedback System.

    PubMed

    Choi, Hailey H; Clark, Jennifer; Jay, Ann K; Filice, Ross W

    2017-08-24

    Feedback is an essential part of medical training, where trainees are provided with information regarding their performance and further directions for improvement. In diagnostic radiology, feedback entails a detailed review of the differences between the residents' preliminary interpretation and the attendings' final interpretation of imaging studies. While the on-call experience of independently interpreting complex cases is important to resident education, the more traditional synchronous "read-out" or joint review is impossible due to multiple constraints. Without an efficient method to compare reports, grade discrepancies, convey salient teaching points, and view images, valuable lessons in image interpretation and report construction are lost. We developed a streamlined web-based system, including report comparison and image viewing, to minimize barriers in asynchronous communication between attending radiologists and on-call residents. Our system provides real-time, end-to-end delivery of case-specific and user-specific feedback in a streamlined, easy-to-view format. We assessed quality improvement subjectively through surveys and objectively through participation metrics. Our web-based feedback system improved user satisfaction for both attending and resident radiologists, and increased attending participation, particularly with regards to cases where substantive discrepancies were identified.

  2. Healing process of the guinea pig common bile duct after end-to-end anastomosis: pathological evaluation after 6 months.

    PubMed

    Zhang, X; Tian, Y; Xu, Z; Wang, L; Hou, C; Ling, X

    2011-01-01

    The healing process of an injured bile duct is always associated with stricture formation, the specific mechanism of which remains unclear. Our objective in this study was to establish a guinea pig common bile duct end-to-end anastomosis (EEA) model after total transection of the common bile duct and to assess the postoperative healing process after 6 months. Thirty-two male guinea pigs were randomly divided into 4 groups with or without operation (a control group without operation and 3 groups examined 2, 3 and 6 months after biliary EEA reconstruction, respectively). Histological, immunohistochemical (proliferating cell nuclear antigen, α-smooth muscle actin, h-caldesmon, and Bcl-2 expression) and ultrastructural characteristics were examined and evaluated. Myofibroblasts (MFBs) were differentiated from smooth muscle cells by α-SMA and h-caldesmon staining. Gross inspection of operated bile ducts revealed coarctation formation at the anastomotic stoma. Histological and immunohistochemical examinations showed significant inflammatory reaction, the presence of MFBs, epithelial proliferation and glandular element hyperplasia. Bcl-2 expression decreased in groups 2 and 3, indicating an epithelial self-protecting mechanism. A guinea pig common bile duct EEA model was successfully established. Glandular elements, bile duct epithelial cells and MFBs all played crucial roles in the bile duct healing process. It seems important to elucidate the pathomechanisms of these components for restoring bile duct physiological function. Copyright © 2011 S. Karger AG, Basel.

  3. Effects of collagen membranes enriched with in vitro-differentiated N1E-115 cells on rat sciatic nerve regeneration after end-to-end repair

    PubMed Central

    2010-01-01

    Peripheral nerves possess the capacity of self-regeneration after traumatic injury but the extent of regeneration is often poor and may benefit from exogenous factors that enhance growth. The use of cellular systems is a rational approach for delivering neurotrophic factors at the nerve lesion site, and in the present study we investigated the effects of enwrapping the site of end-to-end rat sciatic nerve repair with an equine type III collagen membrane enriched or not with N1E-115 pre-differentiated neural cells. After neurotmesis, the sciatic nerve was repaired by end-to-end suture (End-to-End group), end-to-end suture enwrapped with an equine collagen type III membrane (End-to-EndMemb group); and end-to-end suture enwrapped with an equine collagen type III membrane previously covered with neural cells pre-differentiated in vitro from N1E-115 cells (End-to-EndMembCell group). Along the postoperative, motor and sensory functional recovery was evaluated using extensor postural thrust (EPT), withdrawal reflex latency (WRL) and ankle kinematics. After 20 weeks animals were sacrificed and the repaired sciatic nerves were processed for histological and stereological analysis. Results showed that enwrapment of the rapair site with a collagen membrane, with or without neural cell enrichment, did not lead to any significant improvement in most of functional and stereological predictors of nerve regeneration that we have assessed, with the exception of EPT which recovered significantly better after neural cell enriched membrane employment. It can thus be concluded that this particular type of nerve tissue engineering approach has very limited effects on nerve regeneration after sciatic end-to-end nerve reconstruction in the rat. PMID:20149260

  4. Adaptation and validation of a commercial head phantom for cranial radiosurgery dosimetry end-to-end audit.

    PubMed

    Dimitriadis, Alexis; Palmer, Antony L; Thomas, Russell A S; Nisbet, Andrew; Clark, Catharine H

    2017-06-01

    To adapt and validate an anthropomorphic head phantom for use in a cranial radiosurgery audit. Two bespoke inserts were produced for the phantom: one for providing the target and organ at risk for delineation and the other for performing dose measurements. The inserts were tested to assess their positional accuracy. A basic treatment plan dose verification with an ionization chamber was performed to establish a baseline accuracy for the phantom and beam model. The phantom and inserts were then used to perform dose verification measurements of a radiosurgery plan. The dose was measured with alanine pellets, EBT extended dose film and a plastic scintillation detector (PSD). Both inserts showed reproducible positioning (±0.5 mm) and good positional agreement between them (±0.6 mm). The basic treatment plan measurements showed agreement to the treatment planning system (TPS) within 0.5%. Repeated film measurements showed consistent gamma passing rates with good agreement to the TPS. For 2%-2 mm global gamma, the mean passing rate was 96.7% and the variation in passing rates did not exceed 2.1%. The alanine pellets and PSD showed good agreement with the TPS (-0.1% and 0.3% dose difference in the target) and good agreement with each other (within 1%). The adaptations to the phantom showed acceptable accuracies. The presence of alanine and PSD do not affect film measurements significantly, enabling simultaneous measurements by all three detectors. Advances in knowledge: A novel method for thorough end-to-end test of radiosurgery, with capability to incorporate all steps of the clinical pathway in a time-efficient and reproducible manner, suitable for a national audit.

  5. COMPUTATIONAL SIMULATIONS DEMONSTRATE ALTERED WALL SHEAR STRESS IN AORTIC COARCTATION PATIENTS TREATED BY RESECTION WITH END-TO-END ANASTOMOSIS

    PubMed Central

    LaDisa, John F.; Dholakia, Ronak J.; Figueroa, C. Alberto; Vignon-Clementel, Irene E.; Chan, Frandics P.; Samyn, Margaret M.; Cava, Joseph R.; Taylor, Charles A.; Feinstein, Jeffrey A.

    2011-01-01

    Background Atherosclerotic plaque in the descending thoracic aorta (dAo) is related to altered wall shear stress (WSS) for normal patients. Resection with end-to-end anastomosis (RWEA) is the gold standard for coarctation of the aorta (CoA) repair, but may lead to altered WSS indices that contribute to morbidity. Methods Computational fluid dynamics (CFD) models were created from imaging and blood pressure data for control subjects and age- and gender-matched CoA patients treated by RWEA (4 male, 2 female, 15±8 years). CFD analysis incorporated downstream vascular resistance and compliance to generate blood flow velocity, time-averaged WSS (TAWSS) and oscillatory shear index (OSI) results. These indices were quantified longitudinally and circumferentially in the dAo, and several visualization methods were used to highlight regions of potential hemodynamic susceptibility. Results The total dAo area exposed to subnormal TAWSS and OSI was similar between groups, but several statistically significant local differences were revealed. Control subjects experienced left-handed rotating patterns of TAWSS and OSI down the dAo. TAWSS was elevated in CoA patients near the site of residual narrowings and OSI was elevated distally, particularly along the left dAo wall. Differences in WSS indices between groups were negligible more than 5 dAo diameters distal to the aortic arch. Conclusions Localized differences in WSS indices within the dAo of CoA patients treated by RWEA suggest that plaque may form in unique locations influenced by the surgical repair. These regions can be visualized in familiar and intuitive ways allowing clinicians to track their contribution to morbidity in longitudinal studies. PMID:21801315

  6. End-to-end sensor simulation for spectral band selection and optimization with application to the Sentinel-2 mission.

    PubMed

    Segl, Karl; Richter, Rudolf; Küster, Theres; Kaufmann, Hermann

    2012-02-01

    An end-to-end sensor simulation is a proper tool for the prediction of the sensor's performance over a range of conditions that cannot be easily measured. In this study, such a tool has been developed that enables the assessment of the optimum spectral resolution configuration of a sensor based on key applications. It employs the spectral molecular absorption and scattering properties of materials that are used for the identification and determination of the abundances of surface and atmospheric constituents and their interdependence on spatial resolution and signal-to-noise ratio as a basis for the detailed design and consolidation of spectral bands for the future Sentinel-2 sensor. The developed tools allow the computation of synthetic Sentinel-2 spectra that form the frame for the subsequent twofold analysis of bands in the atmospheric absorption and window regions. One part of the study comprises the assessment of optimal spatial and spectral resolution configurations for those bands used for atmospheric correction, optimized with regard to the retrieval of aerosols, water vapor, and the detection of cirrus clouds. The second part of the study presents the optimization of thematic bands, mainly driven by the spectral characteristics of vegetation constituents and minerals. The investigation is performed for different wavelength ranges because most remote sensing applications require the use of specific band combinations rather than single bands. The results from the important "red-edge" and the "short-wave infrared" domains are presented. The recommended optimum spectral design predominantly confirms the sensor parameters given by the European Space Agency. The system is capable of retrieving atmospheric and geobiophysical parameters with enhanced quality compared to existing multispectral sensors. Minor spectral changes of single bands are discussed in the context of typical remote sensing applications, supplemented by the recommendation of a few new bands for

  7. End-to-end gene fusions and their impact on the production of multifunctional biomass degrading enzymes

    SciTech Connect

    Rizk, Mazen; Antranikian, Garabed; Elleuche, Skander

    2012-11-09

    Highlights: Black-Right-Pointing-Pointer Multifunctional enzymes offer an interesting approach for biomass degradation. Black-Right-Pointing-Pointer Size and conformation of separate constructs play a role in the effectiveness of chimeras. Black-Right-Pointing-Pointer A connecting linker allows for maximal flexibility and increased thermostability. Black-Right-Pointing-Pointer Genes with functional similarities are the best choice for fusion candidates. -- Abstract: The reduction of fossil fuels, coupled with its increase in price, has made the search for alternative energy resources more plausible. One of the topics gaining fast interest is the utilization of lignocellulose, the main component of plants. Its primary constituents, cellulose and hemicellulose, can be degraded by a series of enzymes present in microorganisms, into simple sugars, later used for bioethanol production. Thermophilic bacteria have proven to be an interesting source of enzymes required for hydrolysis since they can withstand high and denaturing temperatures, which are usually required for processes involving biomass degradation. However, the cost associated with the whole enzymatic process is staggering. A solution for cost effective and highly active production is through the construction of multifunctional enzyme complexes harboring the function of more than one enzyme needed for the hydrolysis process. There are various strategies for the degradation of complex biomass ranging from the regulation of the enzymes involved, to cellulosomes, and proteins harboring more than one enzymatic activity. In this review, the construction of multifunctional biomass degrading enzymes through end-to-end gene fusions, and its impact on production and activity by choosing the enzymes and linkers is assessed.

  8. Investigating the spatial accuracy of CBCT-guided cranial radiosurgery: A phantom end-to-end test study.

    PubMed

    Calvo-Ortega, Juan-Francisco; Hermida-López, Marcelino; Moragues-Femenía, Sandra; Pozo-Massó, Miquel; Casals-Farran, Joan

    2017-03-01

    To evaluate the spatial accuracy of a frameless cone-beam computed tomography (CBCT)-guided cranial radiosurgery (SRS) using an end-to-end (E2E) phantom test methodology. Five clinical SRS plans were mapped to an acrylic phantom containing a radiochromic film. The resulting phantom-based plans (E2E plans) were delivered four times. The phantom was setup on the treatment table with intentional misalignments, and CBCT-imaging was used to align it prior to E2E plan delivery. Comparisons (global gamma analysis) of the planned and delivered dose to the film were performed using a commercial triple-channel film dosimetry software. The necessary distance-to-agreement to achieve a 95% (DTA95) gamma passing rate for a fixed 3% dose difference provided an estimate of the spatial accuracy of CBCT-guided SRS. Systematic (∑) and random (σ) error components, as well as 95% confidence levels were derived for the DTA95 metric. The overall systematic spatial accuracy averaged over all tests was 1.4mm (SD: 0.2mm), with a corresponding 95% confidence level of 1.8mm. The systematic (Σ) and random (σ) spatial components of the accuracy derived from the E2E tests were 0.2mm and 0.8mm, respectively. The E2E methodology used in this study allowed an estimation of the spatial accuracy of our CBCT-guided SRS procedure. Subsequently, a PTV margin of 2.0mm is currently used in our department. Copyright © 2017 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  9. Greenhouse gas profiling by infrared-laser and microwave occultation in cloudy air: Results from end-to-end simulations

    NASA Astrophysics Data System (ADS)

    Proschek, V.; Kirchengast, G.; Emde, C.; Schweitzer, S.

    2014-11-01

    The new mission concept of microwave and infrared-laser occultation between Low Earth Orbit satellites (LMIO) is capable to provide accurate, consistent, and long-term stable measurements of many essential climate variables. These include temperature, humidity, key greenhouse gases (GHGs) such as carbon dioxide and methane, and line of sight wind speed, all with focus on profiling the upper troposphere and lower stratosphere. The GHG retrieval performance from LMIO data was so far analyzed under clear-air conditions only, without clouds and scintillations from turbulence. Here we present and evaluate an algorithm, built into an already published clear-air algorithm, which copes with cloud and scintillation influences on the infrared-laser transmission profiles used for GHG retrieval. We find that very thin ice clouds fractionally extinct the infrared-laser signals, thicker but broken ice clouds block them over limited altitude ranges, and liquid water clouds generally block them so that their cloud top altitudes typically constitute the limit to tropospheric penetration of profiles. The advanced algorithm penetrates through broken cloudiness. It achieves this by producing a cloud flagging profile from cloud-perturbed infrared-laser signals, which then enables bridging of transmission profile gaps via interpolation. Evaluating the retrieval performance with quasi-realistic end-to-end simulations, including high-resolution cloud data and scintillations from turbulence, we find a small increase only of GHG retrieval RMS errors due to broken-cloud scenes and the profiles remain essentially unbiased as in clear air. These results are encouraging for future LMIO implementation, indicating that GHG profiles can be retrieved through broken cloudiness, maximizing upper troposphere coverage.

  10. Surgical repair of acute Achilles tendon rupture with an end-to-end tendon suture and tendon flap.

    PubMed

    Corradino, B; Di Lorenzo, S; Calamia, C; Moschella, F

    2015-08-01

    Achilles tendon ruptures are becoming more common. Complications after open or minimally invasive surgery are: recurrent rupture (2-8%), wound breakdown, deep infections, granuloma, and fistulas. The authors expose their experience with a personal technique. In 8 patients with acute rupture of Achilles tendon the surgery was performed at least 25 days after trauma. Clinical exam and MR demonstrated in all case a total lesion of tendon. After a posterolateral skin incision the tendon stumps were debrided and suture in end-to-end fashion. A tendon flap was harvested from the proximal part of the tendon, in order to protect and reinforce the suture itself. A plaster cast was applied for 3 weeks and the patients started the rehabilitation protocol. After 4 months all patients returned to pre-injury daily activities. The mean follow up was 13 months (ranged between 6 and 24 months). No major complications occurred. The posterolateral skin incision, not above the tendon, preserves the vascularity of the soft tissues, allows identifying and not accidentally injuring the sural nerve, and prevents the cutaneous scar is overlapped the tendon. In this way is favoured physiological tendon sliding. The preparation of the flap tendon does not weaken the overall strength of the tendon and protects the tendon suture. The tension on sutured stumps is less than being spread over a larger area. In our sample of 8 patients the absence of short-and long-term complications and the rapid functional recovery after surgery suggest that the technique used is safe and effective. Copyright © 2015 Elsevier Ltd. All rights reserved.

  11. Innovative strategy for effective critical laboratory result management: end-to-end process using automation and manual call centre.

    PubMed

    Ti, Lian Kah; Ang, Sophia Bee Leng; Saw, Sharon; Sethi, Sunil Kumar; Yip, James W L

    2012-08-01

    Timely reporting and acknowledgement are crucial steps in critical laboratory results (CLR) management. The authors previously showed that an automated pathway incorporating short messaging system (SMS) texts, auto-escalation, and manual telephone back-up improved the rate and speed of physician acknowledgement compared with manual telephone calling alone. This study investigated if it also improved the rate and speed of physician intervention to CLR and whether utilising the manual back-up affected intervention rates. Data from seven audits between November 2007 and January 2011 were analysed. These audits were carried out to assess the robustness of CLR reporting process in the authors' institution. Comparisons were made in the rate and speed of acknowledgement and intervention between the audits performed before and after automation. Using the automation audits, the authors compared intervention data between communication with SMS only and when manual intervention was required. 1680 CLR were reported during the audit periods. Automation improved the rate (100% vs 84.2%; p<0.001) and speed (median 12 min vs 23 min; p<0.001) of CLR acknowledgement. It also improved the rate (93.7% vs 84.0%, p<0.001) and speed (median 21 min vs 109 min; p<0.001) of CLR intervention. From the automation audits, the use of SMS only did not improve physician intervention rates. The automated communication pathway improved physician intervention rate and time in tandem with improved acknowledgement rate and time when compared with manual telephone calling. The use of manual intervention to augment automation did not adversely affect physician intervention rate, implying that an end-to-end pathway was more important than automation alone.

  12. End-to-end simulations and planning of a small space telescopes: Galaxy Evolution Spectroscopic Explorer: a case study

    NASA Astrophysics Data System (ADS)

    Heap, Sara; Folta, David; Gong, Qian; Howard, Joseph; Hull, Tony; Purves, Lloyd

    2016-08-01

    Large astronomical missions are usually general-purpose telescopes with a suite of instruments optimized for different wavelength regions, spectral resolutions, etc. Their end-to-end (E2E) simulations are typically photons-in to flux-out calculations made to verify that each instrument meets its performance specifications. In contrast, smaller space missions are usually single-purpose telescopes, and their E2E simulations start with the scientific question to be answered and end with an assessment of the effectiveness of the mission in answering the scientific question. Thus, E2E simulations for small missions consist a longer string of calculations than for large missions, as they include not only the telescope and instrumentation, but also the spacecraft, orbit, and external factors such as coordination with other telescopes. Here, we illustrate the strategy and organization of small-mission E2E simulations using the Galaxy Evolution Spectroscopic Explorer (GESE) as a case study. GESE is an Explorer/Probe-class space mission concept with the primary aim of understanding galaxy evolution. Operation of a small survey telescope in space like GESE is usually simpler than operations of large telescopes driven by the varied scientific programs of the observers or by transient events. Nevertheless, both types of telescopes share two common challenges: maximizing the integration time on target, while minimizing operation costs including communication costs and staffing on the ground. We show in the case of GESE how these challenges can be met through a custom orbit and a system design emphasizing simplification and leveraging information from ground-based telescopes.

  13. End-to-end simulations of the visible tunable filter for the Daniel K. Inouye Solar Telescope

    NASA Astrophysics Data System (ADS)

    Schmidt, Wolfgang; Schubert, Matthias; Ellwarth, Monika; Baumgartner, Jörg; Bell, Alexander; Fischer, Andreas; Halbgewachs, Clemens; Heidecke, Frank; Kentischer, Thomas; von der Lühe, Oskar; Scheiffelen, Thomas; Sigwarth, Michael

    2016-08-01

    The Visible Tunable Filter (VTF) is a narrowband tunable filter system for imaging spectroscopy and spectropolarimetry based. The instrument will be one of the first-light instruments of the Daniel K. Inouye Solar Telescope that is currently under construction on Maui (Hawaii). The VTF is being developed by the Kiepenheuer Institut fuer Sonnenphysik in Freiburg as a German contribution to the DKIST. We perform end-to-end simulations of spectropolarimetric observations with the VTF to verify the science requirements of the instrument. The instrument is simulated with two Etalons, and with a single Etalon. The clear aperture of the Etalons is 250 mm, corresponding to a field of view with a diameter of 60 arcsec in the sky (42,000 km on the Sun). To model the large-scale figure errors we employ low-order Zernike polynomials (power and spherical aberration) with amplitudes of 2.5 nm RMS. We use an ideal polarization modulator with equal modulation coefficients of 3-1/2 for the polarization modulation We synthesize Stokes profiles of two iron lines (630.15 nm and 630.25 nm) and for the 854.2 nm line of calcium, for a range of magnetic field values and for several inclination angles. We estimated the photon noise on the basis of the DKIST and VTF transmission values, the atmospheric transmission and the spectral flux from the Sun. For the Fe 630.25 nm line, we obtain a sensitivity of 20 G for the longitudinal component and for 150 G for the transverse component, in agreement with the science requirements for the VTF.

  14. SBSS Demonstrator: A design for efficient demonstration of Space-based Space Surveillance end-to-end capabilities

    NASA Astrophysics Data System (ADS)

    Utzmann, Jens; Flohrer, Tim; Schildknecht, Thomas; Wagner, Axel; Silha, Jiri; Willemsen, Philip; Teston, Frederic

    This paper presents the capabilities of a Space-Based Space Surveillance (SBSS) demonstration mission for Space Surveillance and Tracking (SST) based on a micro-satellite platform. The results have been produced in the frame of ESA’s "Assessment Study for Space Based Space Surveillance Demonstration Mission" performed by the Airbus Defence and Space consortium. Space Surveillance and Tracking is part of Space Situational Awareness (SSA) and covers the detection, tracking and cataloguing of space debris and satellites. Derived SST services comprise a catalogue of these man-made objects, collision warning, detection and characterisation of in-orbit fragmentations, sub-catalogue debris characterisation, etc. The assessment of SBSS in a SST system architecture has shown that both an operational SBSS and also already a well-designed space-based demonstrator can provide substantial performance in terms of surveillance and tracking of beyond-LEO objects. Especially the early deployment of a demonstrator, possible by using standard equipment, could boost initial operating capability and create a self-maintained object catalogue. Furthermore, unique statistical information about small-size LEO debris (mm size) can be collected in-situ. Unlike classical technology demonstration missions, the primary goal is the demonstration and optimisation of the functional elements in a complex end-to-end chain (mission planning, observation strategies, data acquisition, processing and fusion, etc.) until the final products can be offered to the users. Also past and current missions by the US (SBV, SBSS) and Canada (Sapphire, NEOSSat) underline the advantages of space-based space surveillance. The presented SBSS system concept takes the ESA SST System Requirements (derived within the ESA SSA Preparatory Program) into account and aims at fulfilling SST core requirements in a stand-alone manner. Additionally, requirments for detection and characterisation of small-sized LEO debris are

  15. Investigating end-to-end accuracy of image guided radiation treatment delivery using a micro-irradiator

    NASA Astrophysics Data System (ADS)

    Rankine, L. J.; Newton, J.; Bache, S. T.; Das, S. K.; Adamovics, J.; Kirsch, D. G.; Oldham, M.

    2013-11-01

    irradiator was verified to be within 0.5 mm (or 1.0 mm for the 5.0 mm cone) and the cone alignment was verified to be within 0.2 mm (or 0.4 mm for the 1.0 mm cone). The PRESAGE®/DMOS system proved valuable for end-to-end verification of small field IGRT capabilities.

  16. SU-E-J-193: Gamma Knife Perfexion End-To-End Testing with a Commercial Stereotactic Phantom.

    PubMed

    Yount, C; Lyons, K; Rahnema, S; McDonald, D; Peng, J; Vanek, K

    2012-06-01

    To investigate the use of the Lucy ® Stereotactic Phantom (Standard Imaging, Inc.) for Gamma Knife Perfexion radiosurgery quality assurance of the imaging, treatment planning, and dose delivery processes. End-to-end testing of the Perfexion and Gamma Plan version 10.1 has not been previously examined in literature. The phantom was imaged using both the CT and T1- and T2-weighted MR sequences used for treatment planning. For imaging, the isocentric volume insert and fiducial markers were positioned within the phantom. Scans were transferred to the Gamma Plan treatment planning system and were evaluated for geometric and fusion accuracy. A plan was created to deliver 12Gy to the 50% isodose line to the 5.25cm(3) volume. During dose delivery, Gafchromic EBT2 film was positioned in the film insert to replicate the position of the target volume. Dose results were analyzed using RIT software (Radiologic Imaging Technology, Inc.). Image fusion integrity was inspected by overlaying the MR and CT markers (5 fiducial markers spaced 5mm apart) and visually examining the resulting volume insert overlap between the three scans. Geometric accuracy was evaluated by contouring three volumes using Gamma Plan contouring tools. Agreement within 1.1%, 6.7% and 12.2% of the actual volumes was seen with the T1-weighted, T2-weighted, and CT images, respectively. The volume-based acquisition and 1mm slice thickness of the T1-weighted sequence resulted in the most accurate measurement. Geometric measurements along two dimensions showed acceptable accuracy for all imaging modalities within 1.6%. Dosimetry results agreed well with the planned dose. The EBT2 film was calibrated for absolute dose measurements using a dose calibration curve for 0.1-30 Gy and the calibration curve was verified to have <3% error above 1Gy. The Lucy phantom allows for comprehensive quality assurance testing of the Gamma Knife Perfexion radiosurgery process. © 2012 American Association of Physicists in Medicine.

  17. WE-G-BRD-08: End-To-End Targeting Accuracy of the Gamma Knife for Trigeminal Neuralgia

    SciTech Connect

    Brezovich, I; Wu, X; Duan, J; Benhabib, S; Huang, M; Shen, S; Cardan, R; Popple, R

    2014-06-15

    Purpose: Current QA procedures verify accuracy of individual equipment parameters, but may not include CT and MRI localizers. This study uses an end-to-end approach to measure the overall targeting errors in individual patients previously treated for trigeminal neuralgia. Methods: The trigeminal nerve is simulated by a 3 mm long, 3.175 mm (1/8 inch) diameter MRI contrast-filled cavity embedded within a PMMA plastic capsule. The capsule is positioned within the head frame such that the cavity position matches the Gamma Knife coordinates of 10 previously treated patients. Gafchromic EBT2 film is placed at the center of the cavity in coronal and sagittal orientations. The films are marked with a pin prick to identify the cavity center. Treatments are planned for delivery with 4 mm collimators using MRI and CT scans acquired with the clinical localizer boxes and acquisition protocols. Coordinates of shots are chosen so that the cavity is centered within the 50% isodose volume. Following irradiation, the films are scanned and analyzed. Targeting errors are defined as the distance between the pin prick and the centroid of the 50% isodose line. Results: Averaged over 10 patient simulations, targeting errors along the x, y and z coordinates (patient left-to-right, posterior-anterior, head-to-foot) were, respectively, −0.060 +/− 0.363, −0.350 +/− 0.253, and 0.364 +/− 0.191 mm when MRI was used for treatment planning. Planning according to CT exhibited generally smaller errors, namely 0.109 +/− 0.167, −0.191 +/− 0.144, and 0.211 +/− 0.94 mm. The largest errors in MRI and CT planned treatments were, respectively, y = −0.761 and x = 0.428 mm. Conclusion: Unless patient motion or stronger MRI image distortion in actual treatments caused additional errors, all patients received the prescribed dose, i.e., the targeted section of the trig±eminal nerve was contained within the 50% isodose surface in all cases.

  18. Pre-Launch End-to-End Testing Plans for the SPAce Readiness Coherent Lidar Experiment (SPARCLE)

    NASA Technical Reports Server (NTRS)

    Kavaya, Michael J.

    1999-01-01

    The SPAce Readiness Coherent Lidar Experiment (SPARCLE) mission was proposed as a low cost technology demonstration mission, using a 2-micron, 100-mJ, 6-Hz, 25-cm, coherent lidar system based on demonstrated technology. SPARCLE was selected in late October 1997 to be NASA's New Millennium Program (NMP) second earth-observing (EO-2) mission. To maximize the success probability of SPARCLE, NASA/MSFC desired expert guidance in the areas of coherent laser radar (CLR) theory, CLR wind measurement, fielding of CLR systems, CLR alignment validation, and space lidar experience. This led to the formation of the NASA/MSFC Coherent Lidar Technology Advisory Team (CLTAT) in December 1997. A threefold purpose for the advisory team was identified as: 1) guidance to the SPARCLE mission, 2) advice regarding the roadmap of post-SPARCLE coherent Doppler wind lidar (CDWL) space missions and the desired matching technology development plan 3, and 3) general coherent lidar theory, simulation, hardware, and experiment information exchange. The current membership of the CLTAT is shown. Membership does not result in any NASA or other funding at this time. We envision the business of the CLTAT to be conducted mostly by email, teleconference, and occasional meetings. The three meetings of the CLTAT to date, in Jan. 1998, July 1998, and Jan. 1999, have all been collocated with previously scheduled meetings of the Working Group on Space-Based Lidar Winds. The meetings have been very productive. Topics discussed include the SPARCLE technology validation plan including pre-launch end-to-end testing, the space-based wind mission roadmap beyond SPARCLE and its implications on the resultant technology development, the current values and proposed future advancement in lidar system efficiency, and the difference between using single-mode fiber optical mixing vs. the traditional free space optical mixing. attitude information from lidar and non-lidar sensors, and pointing knowledge algorithms will

  19. The End-To-End Safety Verification Process Implemented to Ensure Safe Operations of the Columbus Research Module

    NASA Astrophysics Data System (ADS)

    Arndt, J.; Kreimer, J.

    2010-09-01

    The European Space Laboratory COLUMBUS was launched in February 2008 with NASA Space Shuttle Atlantis. Since successful docking and activation this manned laboratory forms part of the International Space Station(ISS). Depending on the objectives of the Mission Increments the on-orbit configuration of the COLUMBUS Module varies with each increment. This paper describes the end-to-end verification which has been implemented to ensure safe operations under the condition of a changing on-orbit configuration. That verification process has to cover not only the configuration changes as foreseen by the Mission Increment planning but also those configuration changes on short notice which become necessary due to near real-time requests initiated by crew or Flight Control, and changes - most challenging since unpredictable - due to on-orbit anomalies. Subject of the safety verification is on one hand the on orbit configuration itself including the hardware and software products, on the other hand the related Ground facilities needed for commanding of and communication to the on-orbit System. But also the operational products, e.g. the procedures prepared for crew and ground control in accordance to increment planning, are subject of the overall safety verification. In order to analyse the on-orbit configuration for potential hazards and to verify the implementation of the related Safety required hazard controls, a hierarchical approach is applied. The key element of the analytical safety integration of the whole COLUMBUS Payload Complement including hardware owned by International Partners is the Integrated Experiment Hazard Assessment(IEHA). The IEHA especially identifies those hazardous scenarios which could potentially arise through physical and operational interaction of experiments. A major challenge is the implementation of a Safety process which owns quite some rigidity in order to provide reliable verification of on-board Safety and which likewise provides enough

  20. Development and evaluation of an end-to-end test for head and neck IMRT with a novel multiple-dosimetric modality phantom.

    PubMed

    Zakjevskii, Viatcheslav V; Knill, Cory S; Rakowski, Joseph T; Snyder, Michael G

    2016-03-08

    A comprehensive end-to-end test for head and neck IMRT treatments was developed using a custom phantom designed to utilize multiple dosimetry devices. Initial end-to-end test and custom H&N phantom were designed to yield maximum information in anatomical regions significant to H&N plans with respect to: (i) geometric accuracy, (ii) dosimetric accuracy, and (iii) treatment reproducibility. The phantom was designed in collaboration with Integrated Medical Technologies. The phantom was imaged on a CT simulator and the CT was reconstructed with 1 mm slice thickness and imported into Varian's Eclipse treatment planning system. OARs and the PTV were contoured with the aid of Smart Segmentation. A clinical template was used to create an eight-field IMRT plan and dose was calculated with heterogeneity correction on. Plans were delivered with a TrueBeam equipped with a high definition MLC. Preliminary end-to-end results were measured using film, ion chambers, and optically stimulated luminescent dosimeters (OSLDs). Ion chamber dose measurements were compared to the treatment planning system. Films were analyzed with FilmQA Pro using composite gamma index. OSLDs were read with a MicroStar reader using a custom calibration curve. Final phantom design incorporated two axial and one coronal film planes with 18 OSLD locations adjacent to those planes as well as four locations for IMRT ionization chambers below inferior film plane. The end-to-end test was consistently reproducible, resulting in average gamma pass rate greater than 99% using 3%/3 mm analysis criteria, and average OSLD and ion chamber measurements within 1% of planned dose. After initial calibration of OSLD and film systems, the end-to-end test provides next-day results, allowing for integration in routine clinical QA. Preliminary trials have demonstrated that our end-to-end is a reproducible QA tool that enables the ongoing evaluation of dosimetric and geometric accuracy of clinical head and neck treatments.

  1. Measurement of the end-to-end distances between the femoral and tibial insertion sites of the anterior cruciate ligament during knee flexion and with rotational torque.

    PubMed

    Wang, Joon Ho; Kato, Yuki; Ingham, Sheila J M; Maeyama, Akira; Linde-Rosen, Monica; Smolinski, Patrick; Fu, Freddie H

    2012-10-01

    The aim of this study was to determine the end-to-end distance changes in anterior cruciate ligament (ACL) fibers during flexion/extension and internal/external rotation of the knee. The positional relation between the femur and tibia of 10 knees was digitized on a robotic system during flexion/extension and with an internal/external rotational torque (5 Nm). The ACL insertion site data, acquired by 3-dimensional scanning, were superimposed on the positional data. The end-to-end distances of 5 representative points on the femoral and tibial insertion sites of the ACL were calculated. The end-to-end distances of all representative points except the most anterior points were longest at full extension and shortest at 90°. The distances of the anteromedial (AM) and posterolateral (PL) bundles were 37.2 ± 2.1 mm and 27.5 ± 2.8 mm, respectively, at full extension and 34.7 ± 2.4 mm and 20.7 ± 2.3 mm, respectively, at 90°. Only 4 knees had an isometric point, which was 1 of the 3 anterior points. Under an internal torque, both bundles became longer with statistical meaning at all flexion angles (P = .005). The end-to-end distances of all points became longest with internal torque at full extension and shortest with an external torque at 90°. Only 4 of 10 specimens had an isometric point at a variable anterior point. The end-to-end distances of the AM and PL bundles were longer in extension and shorter in flexion. The nonisometric tendency of the ACL and the end-to-end distance change during knee flexion/extension and internal/external rotation should be considered during ACL reconstruction to avoid overconstraint of the graft. Copyright © 2012 Arthroscopy Association of North America. Published by Elsevier Inc. All rights reserved.

  2. SU-E-J-194: Dynamic Tumor Tracking End-To-End Testing Using a 4D Thorax Phantom and EBT3 Films

    SciTech Connect

    Su, Z; Wu, J; Li, Z; Mamalui-Hunter, M

    2014-06-01

    Purpose: To quantify the Vero linac dosimetric accuracy of the tumor dynamic tracking treatment using EBT3 film embedded in a 4D thorax phantom. Methods: A dynamic thorax phantom with tissue equivalent materials and a film insert were used in this study. The thorax phantom was scanned in 4DCT mode with a viscoil embedded in its film insert composed of lung equivalent material. Dynamic tracking planning was performed using the 50% phase CT set with 5 conformal beams at gantry angles of 330, 15, 60, 105 and 150 degrees. Each field is a 3cm by 3cm square centered at viscoil since there was no solid mass target. Total 3 different 1–2cos4 motion profiles were used with varied motion magnitude and cycle frequency. Before treatment plan irradiation, a 4D motion model of the target was established using a series of acquired fluoroscopic images and infrared markers motion positions. During irradiation, fluoroscopic image monitoring viscoil motion was performed to verify model validity. The irradiated films were scanned and the dose maps were compared to the planned Monte Carlo dose distributions. Gamma analyses using 3%–3mm, 2%–3mm, 3%–2mm, 2%–2mm criteria were performed and presented. Results: For each motion pattern, a 4D motion model was built successfully and the target tracking performance was verified with fluoroscopic monitoring of the viscoil motion and its model predicted locations. The film gamma analysis showed the average pass rates among the 3 motion profiles are 98.14%, 96.2%, 91.3% and 85.61% for 3%–3mm, 2%–3mm, 3%–2mm, 2%–2mm criteria. Conclusion: Target dynamic tracking was performed using patient-like breathing patterns in a 4D thorax phantom with EBT3 film insert and a viscoil. There was excellent agreement between acquired and planned dose distributions for all three target motion patterns. This study performed end-to-end testing and verified the treatment accuracy of tumor dynamic tracking.

  3. Thrombosed giant aneurysm of the distal anterior cerebral artery treated with aneurysm resection and proximal pericallosal artery-callosomarginal artery end-to-end anastomosis: Case report and review of the literature.

    PubMed

    Matsushima, Ken; Kawashima, Masatou; Suzuyama, Kenji; Takase, Yukinori; Takao, Tetsuro; Matsushima, Toshio

    2011-01-01

    Giant distal anterior cerebral artery (DACA) aneurysms are extremely rare, with only 32 cases reported in the literature. Most giant DACA aneurysms have features that make standard neck clipping difficult, and bypass surgery is sometimes required, although this surgery was performed in only three reported cases. This report presents the fourth case treated with bypass surgery. A 69-year-old female presented with an unruptured thrombosed giant DACA aneurysm. She underwent wrapping operation 7 years before, but radiological imaging revealed enlargement of the aneurysm at the left pericallosal artery (PerA)-callosomarginal artery (CMA) junction. Before operation, three different strategies were considered for bypass surgery in case the neck could not be clipped. Aneurysm resection and left proximal PerA-CMA end-to-end anastomosis were successfully performed under intraoperative digital subtraction angiography (DSA) and motor-evoked potential (MEP) monitoring. Most DACA aneurysms are located at the PerA-CMA junction. In some cases, adequate retrograde flow to the distal PerA from the posterior or middle cerebral artery can be expected, making distal PerA reconstruction unnecessary. Moreover, when the distal PerA is cut, proximal PerA-CMA end-to-end anastomosis can be easily performed because of reduced tension in both vessels. We therefore conclude that this strategy should be utilized for treating such patients. We also presented here the effectiveness of intraoperative modalities, such as intraoperative DSA and MEP monitoring, for performing a safe operation.

  4. Development and evaluation of an end-to-end test for head and neck IMRT with a novel multiple-dosimetric modality phantom.

    PubMed

    Zakjevskii, Viatcheslav V; Knill, Cory S; Rakowski, Joseph T; Snyder, Michael G

    2016-03-01

    A comprehensive end-to-end test for head and neck IMRT treatments was developed using a custom phantom designed to utilize multiple dosimetry devices. Initial end-to-end test and custom H&N phantom were designed to yield maximum information in anatomical regions significant to H&N plans with respect to: (i) geometric accuracy, (ii) dosimetric accuracy, and (iii) treatment reproducibility. The phantom was designed in collaboration with Integrated Medical Technologies. The phantom was imaged on a CT simulator and the CT was reconstructed with 1 mm slice thickness and imported into Varian's Eclipse treatment planning system. OARs and the PTV were contoured with the aid of Smart Segmentation. A clinical template was used to create an eight-field IMRT plan and dose was calculated with heterogeneity correction on. Plans were delivered with a TrueBeam equipped with a high definition MLC. Preliminary end-to-end results were measured using film, ion chambers, and optically stimulated luminescent dosimeters (OSLDs). Ion chamber dose measurements were compared to the treatment planning system. Films were analyzed with FilmQA Pro using composite gamma index. OSLDs were read with a MicroStar reader using a custom calibration curve. Final phantom design incorporated two axial and one coronal film planes with 18 OSLD locations adjacent to those planes as well as four locations for IMRT ionization chambers below inferior film plane. The end-to-end test was consistently reproducible, resulting in average gamma pass rate greater than 99% using 3%/3 mm analysis criteria, and average OSLD and ion chamber measurements within 1% of planned dose. After initial calibration of OSLD and film systems, the end-to-end test provides next-day results, allowing for integration in routine clinical QA. Preliminary trials have demonstrated that our end-to-end is a reproducible QA tool that enables the ongoing evaluation of dosimetric and geometric accuracy of clinical head and neck treatments. PACS

  5. RTEMP: Exploring an end-to-end, agnostic platform for multidisciplinary real-time analytics in the space physics community and beyond

    NASA Astrophysics Data System (ADS)

    Chaddock, D.; Donovan, E.; Spanswick, E.; Jackel, B. J.

    2014-12-01

    Large-scale, real-time, sensor-driven analytics are a highly effective set of tools in many research environments; however, the barrier to entry is expensive and the learning curve is steep. These systems need to operate efficiently from end to end, with the key aspects being data transmission, acquisition, management and organization, and retrieval. When building a generic multidisciplinary platform, acquisition and data management needs to be designed with scalability and flexibility as the primary focus. Additionally, in order to leverage current sensor web technologies, the integration of common sensor data standards (ie. SensorML and SWE Services) should be supported. Perhaps most important, researchers should be able to get started and integrate the platform into their set of research tools as easily and quickly as possible. The largest issue with current platforms is that the sensor data must be formed and described using the previously mentioned standards. As useful as these standards are for organizing data, they are cumbersome to adopt, often restrictive, and are required to be geospatially-driven. Our solution, RTEMP (Real-time Environment Monitoring Platform), is a real-time analytics platform with over ten years and an estimated two million dollars of investment. It has been developed for our continuously expanding requirements of operating and building remote sensors and supporting equipment for space physics research. A key benefit of our approach is RTEMP's ability to manage agnostic data. This allows data that flows through the system to be structured in any way that best addresses the needs of the sensor operators and data users, enabling extensive flexibility and streamlined development and research. Here we begin with an overview of RTEMP and how it is structured. Additionally, we will showcase the ways that we are using RTEMP and how it is being adopted by researchers in an increasingly broad range of other research fields. We will lay out a

  6. Quantifying residual ionospheric errors in GNSS radio occultation bending angles based on ensembles of profiles from end-to-end simulations

    NASA Astrophysics Data System (ADS)

    Liu, C. L.; Kirchengast, G.; Zhang, K.; Norman, R.; Li, Y.; Zhang, S. C.; Fritzer, J.; Schwaerz, M.; Wu, S. Q.; Tan, Z. X.

    2015-01-01

    The radio occultation (RO) technique using signals from the Global Navigation Satellite System (GNSS), in particular from the Global Positioning System (GPS) so far, is meanwhile widely used to observe the atmosphere for applications such as numerical weather prediction and global climate monitoring. The ionosphere is a major error source in RO measurements at stratospheric altitudes and a linear ionospheric correction of dual-frequency RO bending angles is commonly used to remove the first-order ionospheric effect. However, the residual ionopheric error (RIE) can still be significant so that it needs to be further mitigated for high accuracy applications, especially above about 30 km altitude where the RIE is most relevant compared to the magnitude of the neutral atmospheric bending angle. Quantification and careful analyses for better understanding of the RIE is therefore important towards enabling benchmark-quality stratospheric RO retrievals. Here we present such an analysis of bending angle RIEs covering the stratosphere and mesosphere, using quasi-realistic end-to-end simulations for a full-day ensemble of RO events. Based on the ensemble simulations we assessed the variation of bending angle RIEs, both biases and SDs, with solar activity, latitudinal region, and with or without the assumption of ionospheric spherical symmetry and of co-existing observing system errors. We find that the bending angle RIE biases in the upper stratosphere and mesosphere, and in all latitudinal zones from low- to high-latitudes, have a clear negative tendency and a magnitude increasing with solar activity, in line with recent empirical studies based on real RO data. The maximum RIE biases are found at low latitudes during daytime, where they amount to with in -0.03 to -0.05 μrad, the smallest at high latitudes (0 to -0.01 μrad; quiet space weather and winter conditions). Ionospheric spherical symmetry or asymmetries about the RO event location have only a minor influence on

  7. Quantifying residual ionospheric errors in GNSS radio occultation bending angles based on ensembles of profiles from end-to-end simulations

    NASA Astrophysics Data System (ADS)

    Liu, C. L.; Kirchengast, G.; Zhang, K.; Norman, R.; Li, Y.; Zhang, S. C.; Fritzer, J.; Schwaerz, M.; Wu, S. Q.; Tan, Z. X.

    2015-07-01

    The radio occultation (RO) technique using signals from the Global Navigation Satellite System (GNSS), in particular from the Global Positioning System (GPS) so far, is currently widely used to observe the atmosphere for applications such as numerical weather prediction and global climate monitoring. The ionosphere is a major error source in RO measurements at stratospheric altitudes, and a linear ionospheric correction of dual-frequency RO bending angles is commonly used to remove the first-order ionospheric effect. However, the residual ionospheric error (RIE) can still be significant so that it needs to be further mitigated for high-accuracy applications, especially above about 30 km altitude where the RIE is most relevant compared to the magnitude of the neutral atmospheric bending angle. Quantification and careful analyses for better understanding of the RIE is therefore important for enabling benchmark-quality stratospheric RO retrievals. Here we present such an analysis of bending angle RIEs covering the stratosphere and mesosphere, using quasi-realistic end-to-end simulations for a full-day ensemble of RO events. Based on the ensemble simulations we assessed the variation of bending angle RIEs, both biases and standard deviations, with solar activity, latitudinal region and with or without the assumption of ionospheric spherical symmetry and co-existing observing system errors. We find that the bending angle RIE biases in the upper stratosphere and mesosphere, and in all latitudinal zones from low to high latitudes, have a clear negative tendency and a magnitude increasing with solar activity, which is in line with recent empirical studies based on real RO data although we find smaller bias magnitudes, deserving further study in the future. The maximum RIE biases are found at low latitudes during daytime, where they amount to within -0.03 to -0.05 μrad, the smallest at high latitudes (0 to -0.01 μrad; quiet space weather and winter conditions

  8. Modelling and simulation of the mechanical response of a Dacron graft in the pressurization test and an end-to-end anastomosis.

    PubMed

    Bustos, Claudio A; García-Herrera, Claudio M; Celentano, Diego J

    2016-08-01

    This work presents the modeling and simulation of the mechanical response of a Dacron graft in the pressurization test and its clinical application in the analysis of an end-to-end anastomosis. Both problems are studied via an anisotropic constitutive model that was calibrated by means of previously reported uniaxial tensile tests. First, the simulation of the pressurization test allows the validation of the experimental material characterization that included tests carried out for different levels of axial stretching. Then, the analysis of an end-to-end anastomosis under an idealized geometry is proposed. This case consists in evaluating the mechanical performance of the graft together with the stresses and deformations in the neighborhood of the Dacron with the artery. This research contributes important data to understand the functioning of the graft and the possibility of extending the analysis to complex numerical cases like its insertion in the aortic arch.

  9. End-to-End System Test and Optical Performance Evaluation for the Solar and Heliosphere Observatory (SOHO) Ultraviolet Coronagraph Spectrometer (UVCS)

    NASA Technical Reports Server (NTRS)

    Carosso, Paolo A.; Gardner, Larry D.; Jhabvala, Marzy; Nicolosi, P.

    1997-01-01

    The UVCS is one of the instruments carried by the Solar and Heliospheric Observatory (SOHO), a joint NASA/ESA Spacecraft launched in November 1995. It is designed to perform ultraviolet spectroscopy and visible light polarimetry of the extended solar corona. The primary scientific objectives of the UVCS investigation are to study the physical processes occurring in the extended solar corona, such as: the mechanism of acceleration of the solar wind, the mechanism of coronal plasma heating, the identification of solar wind sources, and the investigation of the plasma properties of the solar wind. The UVCS End-to-End test activities included a comprehensive set of system level functional and optical tests. Although performed under severe schedule constraints, the End-to-End System Test was very successful and served to fully validate the UVCS optical design. All test results showed that the primary scientific objectives of the UVCS Mission were achievable.

  10. End-to-end small bowel anastomosis by temperature controlled CO2 laser soldering and an albumin stent: a feasibility study

    NASA Astrophysics Data System (ADS)

    Simhon, David; Kopelman, Doron; Hashmonai, Moshe; Vasserman, Irena; Dror, Michael; Vasilyev, Tamar; Halpern, Marissa; Kariv, Naam; Katzir, Abraham

    2004-07-01

    Introduction: A feasibility study of small intestinal end to end anastomosis was performed in a rabbit model using temperature controlled CO2 laser system and an albumin stent. Compared with standard suturing or clipping, this method does not introduce foreign materials to the repaired wound and therefore, may lead to better and faster wound healing of the anastomotic site. Methods: Transected rabbits small intestines were either laser soldered using 47% bovine serum albumin and intraluminal albumin stent or served as controls in which conventional continuous two-layer end to end anastomosis was performed manually. The integrity of the anastomosis was investigated at the 14th postoperative day. Results: Postoperative course in both treatments was uneventful. The sutured group presented signs of partial bowel obstruction. Macroscopically, no signs of intraluminal fluid leakage were observed in both treatments. Yet, laser soldered intestinal anastomoses demonstrated significant superiority with respect to adhesions and narrowing of the intestinal lumen. Serial histological examinations revealed better wound healing characteristics of the laser soldered anastomotic site. Conclusion: Laser soldering of intestinal end to end anastomosis provide a faster surgical procedure, compared to standard suture technique, with better wound healing results. It is expected that this technique may be adopted in the future for minimal invasive surgeries.

  11. Direct muscle neurotization after end-to end and end-to-side neurorrhaphy: An experimental study in the rat forelimb model.

    PubMed

    Papalia, Igor; Ronchi, Giulia; Muratori, Luisa; Mazzucco, Alessandra; Magaudda, Ludovico; Geuna, Stefano

    2012-10-15

    The need for the continuous research of new tools for improving motor function recovery after nerve injury is justified by the still often unsatisfactory clinical outcome in these patients. It has been previously shown that the combined use of two reconstructive techniques, namely end-to-side neurorrhaphy and direct muscle neurotization in the rat hindlimb model, can lead to good results in terms of skeletal muscle reinnervation. Here we show that, in the rat forelimb model, the combined use of direct muscle neurotization with either end-to-end or end-to-side neurorrhaphy to reinnervate the denervated flexor digitorum muscles, leads to muscle atrophy prevention over a long postoperative time lapse (10 months). By contrast, very little motor recovery (in case of end-to-end neurorrhaphy) and almost no motor recovery (in case of end-to-side neurorrhaphy) were observed in the grasping activity controlled by flexor digitorum muscles. It can thus be concluded that, at least in the rat, direct muscle neurotization after both end-to-end and end-to-side neurorrhaphy represents a good strategy for preventing denervation-related muscle atrophy but not for regaining the lost motor function.

  12. An integrated healthcare information system for end-to-end standardized exchange and homogeneous management of digital ECG formats.

    PubMed

    Trigo, Jesús Daniel; Martínez, Ignacio; Alesanco, Alvaro; Kollmann, Alexander; Escayola, Javier; Hayn, Dieter; Schreier, Günter; García, José

    2012-07-01

    This paper investigates the application of the enterprise information system (EIS) paradigm to standardized cardiovascular condition monitoring. There are many specifications in cardiology, particularly in the ECG standardization arena. The existence of ECG formats, however, does not guarantee the implementation of homogeneous, standardized solutions for ECG management. In fact, hospital management services need to cope with various ECG formats and, moreover, several different visualization applications. This heterogeneity hampers the normalization of integrated, standardized healthcare information systems, hence the need for finding an appropriate combination of ECG formats and a suitable EIS-based software architecture that enables standardized exchange and homogeneous management of ECG formats. Determining such a combination is one objective of this paper. The second aim is to design and develop the integrated healthcare information system that satisfies the requirements posed by the previous determination. The ECG formats selected include ISO/IEEE11073, Standard Communications Protocol for Computer-Assisted Electrocardiography, and an ECG ontology. The EIS-enabling techniques and technologies selected include web services, simple object access protocol, extensible markup language, or business process execution language. Such a selection ensures the standardized exchange of ECGs within, or across, healthcare information systems while providing modularity and accessibility.

  13. Differentiated CW Policy and Strict Priority Policy for Location-Independent End-to-End Delay in Multi-Hop Wireless Mesh Networks

    NASA Astrophysics Data System (ADS)

    Bae, Yun Han; Kim, Kyung Jae; Park, Jin Soo; Choi, Bong Dae

    We investigate delay analysis of multi-hop wireless mesh network (WMN) where nodes have multi-channel and multiple transceivers to increase the network capacity. The functionality of the multi-channel and multiple transceivers allows the whole WMN to be decomposed into disjoint zones in such a way that i) nodes in a zone are within one-hop distance, and relay node and end nodes with different CWmins contend to access the channel based on IEEE 802.11e EDCA, ii) different channels are assigned to neighbor zones to prevent the hidden node problem, iii) relay nodes can transmit and receive the packets simultaneously by multi-channel and multiple transceivers. With this decomposition of the network, we focus on the delay at a single zone and then the end-to-end delay can be obtained as the sum of zone-delays. In order to have the location-independent end-to-end delay to the gateway regardless of source nodes' locations, we propose two packet management schemes, called the differentiated CW policy and the strict priority policy, at each relay node where relay packets with longer hop count are buffered in higher priority queues according to their experienced hop count. For the differentiated CW policy, a relay node adopts the functionality of IEEE 802.11e EDCA where a higher priority queue has a shorter minimum contention window. We model a typical zone as a one-hop IEEE 802.11e EDCA network under non-saturation condition where priority queues have different packet arrival rates and different minimum contention window sizes. First, we find the PGF (probability generating function) of the HoL-delay of packets at priority queues in a zone. Second, by modeling each queue as M/G/1 queue with the HoL-delay as a service time, we obtain the packet delay (the sum of the queueing delay and the HoL-delay) of each priority queue in a zone. Third, the average end-to-end delay of packet generated at end node in each zone is obtained by summing up the packet delays at each zone. For

  14. Castor (Ricinus communis L.)

    USDA-ARS?s Scientific Manuscript database

    The castor plant has been cultivated for thousands of years, providing a useful source of medicine, lamp fuel and lubricant long before petroleum came into wide use. The oil content of castor seed ranges from 45 to 60%, with most cultivars closer to 50% oil content. The presence of the hydroxy fatty...

  15. An end-to-end system in support of a broad scope of GOES-R sensor and data processing study

    NASA Astrophysics Data System (ADS)

    Huang, Hung-Lung

    2005-08-01

    The mission of NOAA's Geostationary Operational Environmental Satellite System (GOES) R series satellites, in the 2012 time frame, is to provide continuous, near real-time meteorological, oceanographic, solar, and space environment data that supports NOAA's strategic mission goals. It presents an exciting opportunity to explore new instruments, satellite designs, and system architectures utilizing new communication and instrument technologies in order to meet the ever-increasing demands made of Earth observation systems by national agencies and end users alike. The GOES-R sensor suite includes a 16 spectral band Advanced Baseline Imager (ABI), an approximately 1500 high spectral resolution band Hyperspectral Environmental Suite (HES), plus other sensors designed to detect lightning and to explore the ocean, solar and space environment. The Cooperative Institute for Meteorological Satellite Studies (CIMSS) as part of the Space Science and Engineering Center (SSEC) of the University of Wisconsin-Madison, the long time partner of NOAA, has developed the first operational end-to-end processing system for GOES. Based on this heritage, and with recent support from the NASA/NOAA Geosynchrous Imaging FTS (GIFTS) project, the Navy's Multiple University Research Initiative (MURI), and NOAA's GOES-R Risk Reduction program, SSEC has built a near-complete end-to-end system that is capable of simulating sensor measurements from top of atmosphere radiances, raw sensor data (level 0) through calibrated and navigated sensor physical measurements (level 1) to the processed products (level 2). In this paper, the SSEC Hyperspectral Imaging and Sounding Simulator and Processor (HISSP) will be presented in detail. HISSP is capable of demonstrating most of the processing functions such as data compression/decompression, sensor calibration, data processing, algorithm development, and product generation. In summary, HISSP is an end-to-end system designed to support both government and

  16. The Gamma-ray Cherenkov Telescope, an end-to end Schwarzschild-Couder telescope prototype proposed for the Cherenkov Telescope Array

    NASA Astrophysics Data System (ADS)

    Dournaux, J. L.; Abchiche, A.; Allan, D.; Amans, J. P.; Armstrong, T. P.; Balzer, A.; Berge, D.; Boisson, C.; Bousquet, J.-J.; Brown, A. M.; Bryan, M.; Buchholtz, G.; Chadwick, P. M.; Costantini, H.; Cotter, G.; Dangeon, L.; Daniel, M. K.; De Franco, A.; De Frondat, F.; Dumas, D.; Ernenwein, J. P.; Fasola, G.; Funk, S.; Gironnet, J.; Graham, J. A.; Greenshaw, T.; Hameau, B.; Hervet, O.; Hidaka, N.; Hinton, J. A.; Huet, J. M.; Jégouzo, I.; Jogler, T.; Kawashima, T.; Kraush, M.; Lapington, J. S.; Laporte, P.; Lefaucheur, J.; Markoff, S.; Melse, T.; Mohrmann, L.; Molyneux, P.; Nolan, S. J.; Okumura, A.; Osborne, J. P.; Parsons, R. D.; Rosen, S.; Ross, D.; Rowell, G.; Rulten, C. B.; Sato, Y.; Sayède, F.; Schmoll, J.; Schoorlemmer, H.; Servillat, M.; Sol, H.; Stamatescu, V.; Stephan, M.; Stuik, R.; Sykes, J.; Tajima, H.; Thornhill, J.; Tibaldo, L.; Trichard, C.; Vink, J.; Watson, J. J.; White, R.; Yamane, N.; Zech, A.; Zink, A.

    2016-08-01

    The GCT (Gamma-ray Cherenkov Telescope) is a dual-mirror prototype of Small-Sized-Telescopes proposed for the Cherenkov Telescope Array (CTA) and made by an Australian-Dutch-French-German-Indian-Japanese-UK-US consortium. The integration of this end-to-end telescope was achieved in 2015. On-site tests and measurements of the first Cherenkov images on the night sky began on November 2015. This contribution describes the telescope and plans for the pre-production and a large scale production within CTA.

  17. Complications of rectal anastomoses with end-to-end anastomosis (EEA) stapling instrument. Clinical and radiological leak rates and some practical hints.

    PubMed Central

    Dorricott, N. J.; Baddeley, R. M.; Keighley, M. R.; Lee, J.; Oates, G. D.; Alexander-Williams, J.

    1982-01-01

    The complications and results of rectal anastomoses carried out with the end-to-end anastomosis (EEA) stapling instrument on 50 patients by 5 consultant surgeons are recorded. There was a clinical leakage rate of 6% and a radiological leakage rate of 20% assessed by water-soluble contrast enema. The technique has advantages compared with hand-suture by allowing low anastomoses and preservation of sphincters and is accompanied by an acceptably low leakage rate. Despite the cost of disposable cartridges these advantages make the technique economical because of the avoidance of colostomies and reduction in hospital stay. Images FIG. 1 FIG. 2 PMID:7044253

  18. A Vehicle Management End-to-End Testing and Analysis Platform for Validation of Mission and Fault Management Algorithms to Reduce Risk for NASA's Space Launch System

    NASA Technical Reports Server (NTRS)

    Trevino, Luis; Johnson, Stephen B.; Patterson, Jonathan; Teare, David

    2015-01-01

    The development of the Space Launch System (SLS) launch vehicle requires cross discipline teams with extensive knowledge of launch vehicle subsystems, information theory, and autonomous algorithms dealing with all operations from pre-launch through on orbit operations. The characteristics of these systems must be matched with the autonomous algorithm monitoring and mitigation capabilities for accurate control and response to abnormal conditions throughout all vehicle mission flight phases, including precipitating safing actions and crew aborts. This presents a large complex systems engineering challenge being addressed in part by focusing on the specific subsystems handling of off-nominal mission and fault tolerance. Using traditional model based system and software engineering design principles from the Unified Modeling Language (UML), the Mission and Fault Management (M&FM) algorithms are crafted and vetted in specialized Integrated Development Teams composed of multiple development disciplines. NASA also has formed an M&FM team for addressing fault management early in the development lifecycle. This team has developed a dedicated Vehicle Management End-to-End Testbed (VMET) that integrates specific M&FM algorithms, specialized nominal and off-nominal test cases, and vendor-supplied physics-based launch vehicle subsystem models. The flexibility of VMET enables thorough testing of the M&FM algorithms by providing configurable suites of both nominal and off-nominal test cases to validate the algorithms utilizing actual subsystem models. The intent is to validate the algorithms and substantiate them with performance baselines for each of the vehicle subsystems in an independent platform exterior to flight software test processes. In any software development process there is inherent risk in the interpretation and implementation of concepts into software through requirements and test processes. Risk reduction is addressed by working with other organizations such as S

  19. On-Orbit Performance Verification and End-To-End Characterization of the TDRS-H Ka-band Communications Payload

    NASA Technical Reports Server (NTRS)

    Toral, Marco; Wesdock, John; Kassa, Abby; Pogorelc, Patsy; Jenkens, Robert (Technical Monitor)

    2002-01-01

    In June 2000, NASA launched the first of three next generation Tracking and Data Relay Satellites (TDRS-H) equipped with a Ka-band forward and return service capability. This Ka-band service supports forward data rates of up to 25 Mb/sec using the 22.55-23.55 GHz space-to-space allocation. Return services are supported via channel bandwidths of 225 and 650 MHz for data rates up to at least 800 Mb/sec using the 25.25 - 27.5 GHz space-to-space allocation. As part of NASA's acceptance of the TDRS-H spacecraft, an extensive on-orbit calibration, verification and characterization effort was performed to ensure that on-orbit spacecraft performance is within specified limits. This process verified the compliance of the Ka-band communications payload with all performance specifications, and demonstrated an end-to-end Ka-band service capability. This paper summarizes the results of the TDRS-H Ka-band communications payload on-orbit performance verification and end-to-end service characterization. Performance parameters addressed include antenna gain pattern, antenna Gain-to-System Noise Temperature (G/T), Effective Isotropically Radiated Power (EIRP), antenna pointing accuracy, frequency tunability, channel magnitude response, and Ka-band service Bit-Error-Rate (BER) performance.

  20. On-Orbit Performance Verification and End-to-End Characterization of the TDRS-H Ka-Band Communications Payload

    NASA Technical Reports Server (NTRS)

    Toral, Marco; Wesdock, John; Kassa, Abby; Pogorelc, Patsy; Jenkens, Robert (Technical Monitor)

    2002-01-01

    In June 2000, NASA launched the first of three next generation Tracking and Data Relay Satellites (TDRS-H) equipped with a Ka-band forward and return service capability. This Ka-band service supports forward data rates up to 25 Mb/sec using the 22.55 - 23.55 GHz space-to-space allocation. Return services are supported via channel bandwidths of 225 and 650 MHz for data rates up to 800 Mb/sec (QPSK) using the 25.25 - 27.5 GHz space-to-space allocation. As part of NASA's acceptance of the TDRS-H spacecraft, an extensive on-orbit calibration, verification and characterization effort was performed to ensure that on-orbit spacecraft performance is within specified limits. This process verified the compliance of the Ka-band communications payload with all performance specifications and demonstrated an end-to-end Ka-band service capability. This paper summarizes the results of the TDRS-H Ka-band communications payload on-orbit performance verification and end-to-end service characterization. Performance parameters addressed include Effective Isotropically Radiated Power (EIRP), antenna Gain-to-System Noise Temperature (G/T), antenna gain pattern, frequency tunability and accuracy, channel magnitude response, and Ka-band service Bit-Error-Rate (BER) performance.

  1. Comparison of Peripheral Nerve Regeneration with Side-to-side, End-to-side, and End-to-end Repairs: An Experimental Study

    PubMed Central

    Göransson, Harry; Taskinen, Hanna-Stiina; Paavilainen, Pasi; Vahlberg, Tero; Röyttä, Matias

    2016-01-01

    Background: The present study was conducted to find out a tool to enable improved functional recovery with proximal nerve injury. In this experimental study, nerve regeneration was compared between side-to-side (STS), end-to-side (ETS), and end-to-end repairs. Methods: The walk track analysis was used as an outcome of functional recovery. Nerve regeneration was studied with morphometry and histology 6 or 26 weeks postoperatively. Results: All 3 repair techniques showed regeneration of the nerve. From 12 weeks onward, the functional results of the 3 intervention groups were significantly better compared with the unrepaired control group. End-to-end repair was significantly better when compared with the STS and ETS groups. At 26 weeks, the functional and morphometric results and histologic findings did not differ between the STS and ETS groups. The functional results correlated with the morphometric findings in all groups. Conclusions: STS neurorrhaphy showed nerve regeneration, and the end results did not differ from clinically widely used ETS repair. Further studies are warranted to optimize the neurorrhaphy technique and examine possible applications of STS repair in peripheral nerve surgery. PMID:28293523

  2. AAB-sequence living radical chain copolymerization of naturally occurring limonene with maleimide: an end-to-end sequence-regulated copolymer.

    PubMed

    Satoh, Kotaro; Matsuda, Masaru; Nagai, Kanji; Kamigaito, Masami

    2010-07-28

    Sequence control in chain-growth polymerization is still one of the most challenging topics in synthetic polymer chemistry in contrast to natural macromolecules with completely sequence-regulated structures like proteins and DNA. Here, we report the quantitative and highly selective 1:2 sequence-regulated radical copolymerization of naturally occurring (+)-d-limonene (L) and a maleimide (M) in fluoroalcohol giving chiral copolymers with high glass transition temperatures (220-250 degrees C) originating from the specific rigid cyclic structures of the monomers. Furthermore, the combination with a reversible addition-fragmentation chain transfer (RAFT) agent (C-S) via the controlled/living radical polymerization resulted in end-to-end sequence-regulated copolymers [C-(M-M-L)(n)-M-S] with both highly sequenced chain ends and main-chain repeating units as well as controlled molecular weights.

  3. Demonstration of a fully-coupled end-to-end model for small pelagic fish using sardine and anchovy in the California Current

    NASA Astrophysics Data System (ADS)

    Rose, Kenneth A.; Fiechter, Jerome; Curchitser, Enrique N.; Hedstrom, Kate; Bernal, Miguel; Creekmore, Sean; Haynie, Alan; Ito, Shin-ichi; Lluch-Cota, Salvador; Megrey, Bernard A.; Edwards, Chris A.; Checkley, Dave; Koslow, Tony; McClatchie, Sam; Werner, Francisco; MacCall, Alec; Agostini, Vera

    2015-11-01

    We describe and document an end-to-end model of anchovy and sardine population dynamics in the California Current as a proof of principle that such coupled models can be developed and implemented. The end-to-end model is 3-dimensional, time-varying, and multispecies, and consists of four coupled submodels: hydrodynamics, Eulerian nutrient-phytoplankton-zooplankton (NPZ), an individual-based full life cycle anchovy and sardine submodel, and an agent-based fishing fleet submodel. A predator roughly mimicking albacore was included as individuals that consumed anchovy and sardine. All submodels were coded within the ROMS open-source community model, and used the same resolution spatial grid and were all solved simultaneously to allow for possible feedbacks among the submodels. We used a super-individual approach and solved the coupled models on a distributed memory parallel computer, both of which created challenging but resolvable bookkeeping challenges. The anchovy and sardine growth, mortality, reproduction, and movement, and the fishing fleet submodel, were each calibrated using simplified grids before being inserted into the full end-to-end model. An historical simulation of 1959-2008 was performed, and the latter 45 years analyzed. Sea surface height (SSH) and sea surface temperature (SST) for the historical simulation showed strong horizontal gradients and multi-year scale temporal oscillations related to various climate indices (PDO, NPGO), and both showed responses to ENSO variability. Simulated total phytoplankton was lower during strong El Nino events and higher for the strong 1999 La Nina event. The three zooplankton groups generally corresponded to the spatial and temporal variation in simulated total phytoplankton. Simulated biomasses of anchovy and sardine were within the historical range of observed biomasses but predicted biomasses showed much less inter-annual variation. Anomalies of annual biomasses of anchovy and sardine showed a switch in the mid

  4. Mixed integer nonlinear programming model of wireless pricing scheme with QoS attribute of bandwidth and end-to-end delay

    NASA Astrophysics Data System (ADS)

    Irmeilyana, Puspita, Fitri Maya; Indrawati

    2016-02-01

    The pricing for wireless networks is developed by considering linearity factors, elasticity price and price factors. Mixed Integer Nonlinear Programming of wireless pricing model is proposed as the nonlinear programming problem that can be solved optimally using LINGO 13.0. The solutions are expected to give some information about the connections between the acceptance factor and the price. Previous model worked on the model that focuses on bandwidth as the QoS attribute. The models attempt to maximize the total price for a connection based on QoS parameter. The QoS attributes used will be the bandwidth and the end to end delay that affect the traffic. The maximum goal to maximum price is achieved when the provider determine the requirement for the increment or decrement of price change due to QoS change and amount of QoS value.

  5. End-to-End Study of the Transfer of Energy from Magnetosheath Ion Precipitation to the Ionospheric Cusp and Resulting Ion Outflow to the Magnetosphere

    NASA Technical Reports Server (NTRS)

    Coffey, Victoria; Chandler, Michael; Singh, Nagendra; Avanov, Levon

    2003-01-01

    We will show results from an end-to-end study of the energy transfer from injected magnetosheath plasmas to the near-Earth magnetospheric and ionospheric plasmas and the resulting ion outflow to the magnetosphere. This study includes modeling of the evolution of the magnetosheath precipitation in the cusp using a kinetic code with a realistic magnetic field configuration. These evolved, highly non-Maxwellian distributions are used as input to a 2D PIC code to analyze the resulting wave generation. The wave analysis is used in the kinetic code as input to the cold ionospheric ions to study the transfer of energy to these ions and their outflow to the magnetosphere. Observations from the Thermal Ion Dynamics Experiment (TIDE) and other instruments on the Polar Spacecraft will be compared to the modeling.

  6. End-to-End Study of the Transfer of Energy from Magnetosheath Ion Precipitation to the Ionospheric Cusp and Resulting Ion Outflow to the Magnetosphere

    NASA Technical Reports Server (NTRS)

    Coffey, Victoria; Chandler, Michael; Singh, Nagendra; Avanov, Levon

    2003-01-01

    We will show results from an end-to-end study of the energy transfer from injected magnetosheath plasmas to the near-Earth magnetospheric and ionospheric plasmas and the resulting ion outflow to the magnetosphere. This study includes modeling of the evolution of the magnetosheath precipitation in the cusp using a kinetic code with a realistic magnetic field configuration. These evolved, highly non-Maxwellian distributions are used as input to a 2D PIC code to analyze the resulting wave generation. The wave analysis is used in the kinetic code as input to the cold ionospheric ions to study the transfer of energy to these ions and their outflow to the magnetosphere. Observations from the Thermal Ion Dynamics Experiment (TIDE) and other instruments on the Polar Spacecraft will be compared to the modeling.

  7. The role of environmental controls in determining sardine and anchovy population cycles in the California Current: Analysis of an end-to-end model

    NASA Astrophysics Data System (ADS)

    Fiechter, Jerome; Rose, Kenneth A.; Curchitser, Enrique N.; Hedstrom, Katherine S.

    2015-11-01

    Sardine and anchovy are two forage species of particular interest because of their low-frequency cycles in adult abundance in boundary current regions, combined with a commercially relevant contribution to the global marine food catch. While several hypotheses have been put forth to explain decadal shifts in sardine and anchovy populations, a mechanistic basis for how the physics, biogeochemistry, and biology combine to produce patterns of synchronous variability across widely separated systems has remained elusive. The present study uses a 50-year (1959-2008) simulation of a fully coupled end-to-end ecosystem model configured for sardine and anchovy in the California Current System to investigate how environmental processes control their population dynamics. The results illustrate that slightly different temperature and diet preferences can lead to significantly different responses to environmental variability. Simulated adult population fluctuations are associated with age-1 growth (via age-2 egg production) and prey availability for anchovy, while they depend primarily on age-0 survival and temperature for sardine. The analysis also hints at potential linkages to known modes of climate variability, whereby changes in adult abundance are related to ENSO for anchovy and to the PDO for sardine. The connection to the PDO and ENSO is consistent with modes of interannual and decadal variability that would alternatively favor anchovy during years of cooler temperatures and higher prey availability, and sardine during years of warmer temperatures and lower prey availability. While the end-to-end ecosystem model provides valuable insight on potential relationships between environmental conditions and sardine and anchovy population dynamics, understanding the complex interplay, and potential lags, between the full array of processes controlling their abundances in the California Current System remains an on-going challenge.

  8. A Vehicle Management End-to-End Testing and Analysis Platform for Validation of Mission and Fault Management Algorithms to Reduce Risk for NASAs Space Launch System

    NASA Technical Reports Server (NTRS)

    Trevino, Luis; Johnson, Stephen B.; Patterson, Jonathan; Teare, David

    2015-01-01

    The engineering development of the National Aeronautics and Space Administration's (NASA) new Space Launch System (SLS) requires cross discipline teams with extensive knowledge of launch vehicle subsystems, information theory, and autonomous algorithms dealing with all operations from pre-launch through on orbit operations. The nominal and off-nominal characteristics of SLS's elements and subsystems must be understood and matched with the autonomous algorithm monitoring and mitigation capabilities for accurate control and response to abnormal conditions throughout all vehicle mission flight phases, including precipitating safing actions and crew aborts. This presents a large and complex systems engineering challenge, which is being addressed in part by focusing on the specific subsystems involved in the handling of off-nominal mission and fault tolerance with response management. Using traditional model-based system and software engineering design principles from the Unified Modeling Language (UML) and Systems Modeling Language (SysML), the Mission and Fault Management (M&FM) algorithms for the vehicle are crafted and vetted in Integrated Development Teams (IDTs) composed of multiple development disciplines such as Systems Engineering (SE), Flight Software (FSW), Safety and Mission Assurance (S&MA) and the major subsystems and vehicle elements such as Main Propulsion Systems (MPS), boosters, avionics, Guidance, Navigation, and Control (GNC), Thrust Vector Control (TVC), and liquid engines. These model-based algorithms and their development lifecycle from inception through FSW certification are an important focus of SLS's development effort to further ensure reliable detection and response to off-nominal vehicle states during all phases of vehicle operation from pre-launch through end of flight. To test and validate these M&FM algorithms a dedicated test-bed was developed for full Vehicle Management End-to-End Testing (VMET). For addressing fault management (FM

  9. A Vehicle Management End-to-End Testing and Analysis Platform for Validation of Mission and Fault Management Algorithms to Reduce Risk for NASA's Space Launch System

    NASA Technical Reports Server (NTRS)

    Trevino, Luis; Patterson, Jonathan; Teare, David; Johnson, Stephen

    2015-01-01

    integrates specific M&FM algorithms, specialized nominal and off-nominal test cases, and vendor-supplied physics-based launch vehicle subsystem models. Additionally, the team has developed processes for implementing and validating these algorithms for concept validation and risk reduction for the SLS program. The flexibility of the Vehicle Management End-to-end Testbed (VMET) enables thorough testing of the M&FM algorithms by providing configurable suites of both nominal and off-nominal test cases to validate the developed algorithms utilizing actual subsystem models such as MPS. The intent of VMET is to validate the M&FM algorithms and substantiate them with performance baselines for each of the target vehicle subsystems in an independent platform exterior to the flight software development infrastructure and its related testing entities. In any software development process there is inherent risk in the interpretation and implementation of concepts into software through requirements and test cases into flight software compounded with potential human errors throughout the development lifecycle. Risk reduction is addressed by the M&FM analysis group working with other organizations such as S&MA, Structures and Environments, GNC, Orion, the Crew Office, Flight Operations, and Ground Operations by assessing performance of the M&FM algorithms in terms of their ability to reduce Loss of Mission and Loss of Crew probabilities. In addition, through state machine and diagnostic modeling, analysis efforts investigate a broader suite of failure effects and associated detection and responses that can be tested in VMET to ensure that failures can be detected, and confirm that responses do not create additional risks or cause undesired states through interactive dynamic effects with other algorithms and systems. VMET further contributes to risk reduction by prototyping and exercising the M&FM algorithms early in their implementation and without any inherent hindrances such as meeting FSW

  10. Imaging and dosimetric errors in 4D PET/CT-guided radiotherapy from patient-specific respiratory patterns: a dynamic motion phantom end-to-end study

    NASA Astrophysics Data System (ADS)

    Bowen, S. R.; Nyflot, M. J.; Herrmann, C.; Groh, C. M.; Meyer, J.; Wollenweber, S. D.; Stearns, C. W.; Kinahan, P. E.; Sandison, G. A.

    2015-05-01

    Effective positron emission tomography / computed tomography (PET/CT) guidance in radiotherapy of lung cancer requires estimation and mitigation of errors due to respiratory motion. An end-to-end workflow was developed to measure patient-specific motion-induced uncertainties in imaging, treatment planning, and radiation delivery with respiratory motion phantoms and dosimeters. A custom torso phantom with inserts mimicking normal lung tissue and lung lesion was filled with [18F]FDG. The lung lesion insert was driven by six different patient-specific respiratory patterns or kept stationary. PET/CT images were acquired under motionless ground truth, tidal breathing motion-averaged (3D), and respiratory phase-correlated (4D) conditions. Target volumes were estimated by standardized uptake value (SUV) thresholds that accurately defined the ground-truth lesion volume. Non-uniform dose-painting plans using volumetrically modulated arc therapy were optimized for fixed normal lung and spinal cord objectives and variable PET-based target objectives. Resulting plans were delivered to a cylindrical diode array at rest, in motion on a platform driven by the same respiratory patterns (3D), or motion-compensated by a robotic couch with an infrared camera tracking system (4D). Errors were estimated relative to the static ground truth condition for mean target-to-background (T/Bmean) ratios, target volumes, planned equivalent uniform target doses, and 2%-2 mm gamma delivery passing rates. Relative to motionless ground truth conditions, PET/CT imaging errors were on the order of 10-20%, treatment planning errors were 5-10%, and treatment delivery errors were 5-30% without motion compensation. Errors from residual motion following compensation methods were reduced to 5-10% in PET/CT imaging, <5% in treatment planning, and <2% in treatment delivery. We have demonstrated that estimation of respiratory motion uncertainty and its propagation from PET/CT imaging to RT planning, and RT

  11. Imaging and dosimetric errors in 4D PET/CT-guided radiotherapy from patient-specific respiratory patterns: a dynamic motion phantom end-to-end study.

    PubMed

    Bowen, S R; Nyflot, M J; Herrmann, C; Groh, C M; Meyer, J; Wollenweber, S D; Stearns, C W; Kinahan, P E; Sandison, G A

    2015-05-07

    Effective positron emission tomography / computed tomography (PET/CT) guidance in radiotherapy of lung cancer requires estimation and mitigation of errors due to respiratory motion. An end-to-end workflow was developed to measure patient-specific motion-induced uncertainties in imaging, treatment planning, and radiation delivery with respiratory motion phantoms and dosimeters. A custom torso phantom with inserts mimicking normal lung tissue and lung lesion was filled with [(18)F]FDG. The lung lesion insert was driven by six different patient-specific respiratory patterns or kept stationary. PET/CT images were acquired under motionless ground truth, tidal breathing motion-averaged (3D), and respiratory phase-correlated (4D) conditions. Target volumes were estimated by standardized uptake value (SUV) thresholds that accurately defined the ground-truth lesion volume. Non-uniform dose-painting plans using volumetrically modulated arc therapy were optimized for fixed normal lung and spinal cord objectives and variable PET-based target objectives. Resulting plans were delivered to a cylindrical diode array at rest, in motion on a platform driven by the same respiratory patterns (3D), or motion-compensated by a robotic couch with an infrared camera tracking system (4D). Errors were estimated relative to the static ground truth condition for mean target-to-background (T/Bmean) ratios, target volumes, planned equivalent uniform target doses, and 2%-2 mm gamma delivery passing rates. Relative to motionless ground truth conditions, PET/CT imaging errors were on the order of 10-20%, treatment planning errors were 5-10%, and treatment delivery errors were 5-30% without motion compensation. Errors from residual motion following compensation methods were reduced to 5-10% in PET/CT imaging, <5% in treatment planning, and <2% in treatment delivery. We have demonstrated that estimation of respiratory motion uncertainty and its propagation from PET/CT imaging to RT planning, and

  12. Imaging and dosimetric errors in 4D PET/CT-guided radiotherapy from patient-specific respiratory patterns: a dynamic motion phantom end-to-end study

    PubMed Central

    Bowen, S R; Nyflot, M J; Hermann, C; Groh, C; Meyer, J; Wollenweber, S D; Stearns, C W; Kinahan, P E; Sandison, G A

    2015-01-01

    Effective positron emission tomography/computed tomography (PET/CT) guidance in radiotherapy of lung cancer requires estimation and mitigation of errors due to respiratory motion. An end-to-end workflow was developed to measure patient-specific motion-induced uncertainties in imaging, treatment planning, and radiation delivery with respiratory motion phantoms and dosimeters. A custom torso phantom with inserts mimicking normal lung tissue and lung lesion was filled with [18F]FDG. The lung lesion insert was driven by 6 different patient-specific respiratory patterns or kept stationary. PET/CT images were acquired under motionless ground truth, tidal breathing motion-averaged (3D), and respiratory phase-correlated (4D) conditions. Target volumes were estimated by standardized uptake value (SUV) thresholds that accurately defined the ground-truth lesion volume. Non-uniform dose-painting plans using volumetrically modulated arc therapy (VMAT) were optimized for fixed normal lung and spinal cord objectives and variable PET-based target objectives. Resulting plans were delivered to a cylindrical diode array at rest, in motion on a platform driven by the same respiratory patterns (3D), or motion-compensated by a robotic couch with an infrared camera tracking system (4D). Errors were estimated relative to the static ground truth condition for mean target-to-background (T/Bmean) ratios, target volumes, planned equivalent uniform target doses (EUD), and 2%-2mm gamma delivery passing rates. Relative to motionless ground truth conditions, PET/CT imaging errors were on the order of 10–20%, treatment planning errors were 5–10%, and treatment delivery errors were 5–30% without motion compensation. Errors from residual motion following compensation methods were reduced to 5–10% in PET/CT imaging, < 5% in treatment planning, and < 2% in treatment delivery. We have demonstrated that estimation of respiratory motion uncertainty and its propagation from PET/CT imaging to RT

  13. End-to-end ureteral anastomosis and double-pigtail ureteral stent placement for treatment of iatrogenic ureteral trauma in two dogs.

    PubMed

    Wormser, Chloe; Clarke, Dana L; Aronson, Lillian R

    2015-07-01

    A 6-month-old spayed female Soft-Coated Wheaten Terrier and 8-month-old spayed female Shih Tzu were referred because of complications related to inadvertent ureteral ligation and transection during recent ovariohysterectomy. The Soft-Coated Wheaten Terrier had a 2-day history of stranguria and polyuria that began after ovariohysterectomy. Initial examination findings were unremarkable with the exception of high rectal temperature. The Shih Tzu had a 10-day history of pyrexia, vomiting, diarrhea, and stranguria that began after ovariohysterectomy. On examination, the dog had signs of depression; clinicopathologic tests revealed hypoalbuminemia, neutrophilia, lymphocytosis, and monocytosis. Abdominal ultrasonography was performed for both dogs, revealing severe unilateral pyelectasia and hydroureter (proximal portion). Both dogs underwent exploratory celiotomy; ureteral ligation and transection was confirmed. Ventral cystotomy was performed to allow retrograde placement of a double-pigtail ureteral stent into the affected ureter and renal pelvis. End-to-end ureteral anastomosis was performed over the stent with the aid of an operating microscope. Stent position was confirmed via fluoroscopy, and incisions were closed routinely. Dogs continued to have intermittent signs of stranguria until stent removal via cystoscopy 6 or 7 weeks after surgery. Ultrasonographic examination of the urogenital tract was performed 2 or 4 months after surgery, revealing resolution of pyelectasia and hydroureter. The surgical technique used provided a viable option for preserving renal function in dogs with focal, iatrogenic ureteral trauma. Use of a ureteral stent facilitated ureteral anastomosis and minimized postoperative complications.

  14. End-to-end simulation of a K-band LEO-LEO satellite link for estimating water vapor in the low troposphere

    NASA Astrophysics Data System (ADS)

    Facheris, Luca; Cuccoli, Fabrizio; Argenti, Fabrizio

    2004-11-01

    A new differential measurement concept is presented for retrieving the total content of water vapor (IWV, Integrated Water Vapor) along the propagation path between two Low Earth Orbiting (LEO) satellites, while such path is immersing in the atmosphere during a so called set occultation. The new approach, referred to as DSA (Differential Spectral Absorption) method, is based on the simultaneous measurement of the total attenuation at two relatively close frequencies in the K band, and on the estimate of a "spectral sensitivity parameter" that is highly correlated to the IWV content of the LEO-LEO link in the low troposphere. The DSA approach has the potential to overcome all spectrally 'flat' and spectrally correlated phenomena (atmospheric scintillation among these) and provides estimates that can then be usefully integrated with standard radio occultation data products. In the paper we describe the signaling structure chosen for DSA measurements and the transmit-receive system used to simulate an end-to-end transmission during a complete LEO-LEO set occultation. Simulations are based on atmospheric models and on real radiosonde data, which allows us to account for the natural variability of the atmospheric conditions. The effects on the IWV estimates of impairments such as thermal noise at the receiver, atmospheric scintillation, multipath and defocusing are evaluated.

  15. Performances of the fractal iterative method with an internal model control law on the ESO end-to-end ELT adaptive optics simulator

    NASA Astrophysics Data System (ADS)

    Béchet, C.; Le Louarn, M.; Tallon, M.; Thiébaut, É.

    2008-07-01

    Adaptive Optics systems under study for the Extremely Large Telescopes gave rise to a new generation of algorithms for both wavefront reconstruction and the control law. In the first place, the large number of controlled actuators impose the use of computationally efficient methods. Secondly, the performance criterion is no longer solely based on nulling residual measurements. Priors on turbulence must be inserted. In order to satisfy these two requirements, we suggested to associate the Fractal Iterative Method for the estimation step with an Internal Model Control. This combination has now been tested on an end-to-end adaptive optics numerical simulator at ESO, named Octopus. Results are presented here and performance of our method is compared to the classical Matrix-Vector Multiplication combined with a pure integrator. In the light of a theoretical analysis of our control algorithm, we investigate the influence of several errors contributions on our simulations. The reconstruction error varies with the signal-to-noise ratio but is limited by the use of priors. The ratio between the system loop delay and the wavefront coherence time also impacts on the reachable Strehl ratio. Whereas no instabilities are observed, correction quality is obviously affected at low flux, when subapertures extinctions are frequent. Last but not least, the simulations have demonstrated the robustness of the method with respect to sensor modeling errors and actuators misalignments.

  16. An End-to-End Trainable Neural Network for Image-based Sequence Recognition and Its Application to Scene Text Recognition.

    PubMed

    Shi, Baoguang; Bai, Xiang; Yao, Cong

    2016-12-29

    Image-based sequence recognition has been a long-standing research topic in computer vision. In this paper, we investigate the problem of scene text recognition, which is among the most important and challenging tasks in image-based sequence recognition. A novel neural network architecture, which integrates feature extraction, sequence modeling and transcription into a unified framework, is proposed. Compared with previous systems for scene text recognition, the proposed architecture possesses four distinctive properties: (1) It is end-to-end trainable, in contrast to most of the existing algorithms whose components are separately trained and tuned. (2) It naturally handles sequences in arbitrary lengths, involving no character segmentation or horizontal scale normalization. (3) It is not confined to any predefined lexicon and achieves remarkable performances in both lexicon-free and lexicon-based scene text recognition tasks. (4) It generates an effective yet much smaller model, which is more practical for real-world application scenarios. The experiments on standard benchmarks, including the IIIT-5K, Street View Text and ICDAR datasets, demonstrate the superiority of the proposed algorithm over the prior arts. Moreover, the proposed algorithm performs well in the task of image-based music score recognition, which evidently verifies the generality of it.

  17. Land Mobile Satellite Service (LMSS) channel simulator: An end-to-end hardware simulation and study of the LMSS communications links

    NASA Technical Reports Server (NTRS)

    Salmasi, A. B. (Editor); Springett, J. C.; Sumida, J. T.; Richter, P. H.

    1984-01-01

    The design and implementation of the Land Mobile Satellite Service (LMSS) channel simulator as a facility for an end to end hardware simulation of the LMSS communications links, primarily with the mobile terminal is described. A number of studies are reported which show the applications of the channel simulator as a facility for validation and assessment of the LMSS design requirements and capabilities by performing quantitative measurements and qualitative audio evaluations for various link design parameters and channel impairments under simulated LMSS operating conditions. As a first application, the LMSS channel simulator was used in the evaluation of a system based on the voice processing and modulation (e.g., NBFM with 30 kHz of channel spacing and a 2 kHz rms frequency deviation for average talkers) selected for the Bell System's Advanced Mobile Phone Service (AMPS). The various details of the hardware design, qualitative audio evaluation techniques, signal to channel impairment measurement techniques, the justifications for criteria of different parameter selection in regards to the voice processing and modulation methods, and the results of a number of parametric studies are further described.

  18. Overview of Non-nuclear Testing of the Safe, Affordable 30-kW Fission Engine, Including End-to-End Demonstrator Testing

    NASA Technical Reports Server (NTRS)

    VanDyke, M. K.; Martin, J. J.; Houts, M. G.

    2003-01-01

    Successful development of space fission systems will require an extensive program of affordable and realistic testing. In addition to tests related to design/development of the fission system, realistic testing of the actual flight unit must also be performed. At the power levels under consideration (3-300 kW electric power), almost all technical issues are thermal or stress related and will not be strongly affected by the radiation environment. These issues can be resolved more thoroughly, less expensively, and in a more timely fashing with nonnuclear testing, provided it is prototypic of the system in question. This approach was used for the safe, affordable fission engine test article development program and accomplished viz cooperative efforts with Department of Energy labs, industry, universiites, and other NASA centers. This Technical Memorandum covers the analysis, testing, and data reduction of a 30-kW simulated reactor as well as an end-to-end demonstrator, including a power conversion system and an electric propulsion engine, the first of its kind in the United States.

  19. Determination of anions using monolithic capillary column ion chromatography with end-to-end differential contactless conductometric detectors under resonance approach.

    PubMed

    Zhang, Zhenli; Li, Dongdong; Liu, Xueyong; Subhani, Qamar; Zhu, Yan; Kang, Qi; Shen, Dazhong

    2012-06-21

    An end-to-end differential measurement approach with capacitively coupled contactless conductivity detection (C(4)D) was applied to anion-exchange monolithic capillary column ion chromatography. The column was prepared by thermally initiated radical polymerization of poly(glycidyl methacrylate) in a fused-silica capillary of 320 μm i.d. and modified by quaternary ammonium latex surface coating. Two C(4)Ds were placed near both ends of the capillary column and the output difference between them was measured. With 15 mM potassium hydrogen phthalate used as the eluent, good separation of a mixture of inorganic anions (F(-), Cl(-), NO(2)(-), NO(3)(-)) was achieved. The detection limits of conventional C(4)D are 1.6, 0.28, 0.53, and 0.47 mg L(-1) for F(-), Cl(-), NO(2)(-), and NO(3)(-), respectively. To further enhance the sensitivity, the capacitive impedance from C(4)D was neutralized by an inductive impedance from a piezoelectric resonator. An increase in sensitivity by a factor of 7-8 was achieved in the resonating C(4)D in comparison with the conventional C(4)D. The detection limits of the resonating C(4)D are 0.23, 0.041, 0.065, and 0.059 mg L(-1) for F(-), Cl(-), NO(2)(-), and NO(3)(-), respectively. The response of the resonating C(4)D was analyzed based on an equivalent circuit model.

  20. 46 CFR 148.235 - Castor beans.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 46 Shipping 5 2012-10-01 2012-10-01 false Castor beans. 148.235 Section 148.235 Shipping COAST... THAT REQUIRE SPECIAL HANDLING Special Requirements for Certain Materials § 148.235 Castor beans. (a) This part applies only to the stowage and transportation of whole castor beans. Castor meal, castor...

  1. 46 CFR 148.235 - Castor beans.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 46 Shipping 5 2014-10-01 2014-10-01 false Castor beans. 148.235 Section 148.235 Shipping COAST... THAT REQUIRE SPECIAL HANDLING Special Requirements for Certain Materials § 148.235 Castor beans. (a) This part applies only to the stowage and transportation of whole castor beans. Castor meal, castor...

  2. 46 CFR 148.235 - Castor beans.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 46 Shipping 5 2013-10-01 2013-10-01 false Castor beans. 148.235 Section 148.235 Shipping COAST... THAT REQUIRE SPECIAL HANDLING Special Requirements for Certain Materials § 148.235 Castor beans. (a) This part applies only to the stowage and transportation of whole castor beans. Castor meal, castor...

  3. 46 CFR 148.235 - Castor beans.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 46 Shipping 5 2011-10-01 2011-10-01 false Castor beans. 148.235 Section 148.235 Shipping COAST... THAT REQUIRE SPECIAL HANDLING Special Requirements for Certain Materials § 148.235 Castor beans. (a) This part applies only to the stowage and transportation of whole castor beans. Castor meal, castor...

  4. Experimental demonstration of a record high 11.25Gb/s real-time optical OFDM transceiver supporting 25km SMF end-to-end transmission in simple IMDD systems.

    PubMed

    Giddings, R P; Jin, X Q; Hugues-Salas, E; Giacoumidis, E; Wei, J L; Tang, J M

    2010-03-15

    The fastest ever 11.25Gb/s real-time FPGA-based optical orthogonal frequency division multiplexing (OOFDM) transceivers utilizing 64-QAM encoding/decoding and significantly improved variable power loading are experimentally demonstrated, for the first time, incorporating advanced functionalities of on-line performance monitoring, live system parameter optimization and channel estimation. Real-time end-to-end transmission of an 11.25Gb/s 64-QAM-encoded OOFDM signal with a high electrical spectral efficiency of 5.625bit/s/Hz over 25km of standard and MetroCor single-mode fibres is successfully achieved with respective power penalties of 0.3dB and -0.2dB at a BER of 1.0 x 10(-3) in a directly modulated DFB laser-based intensity modulation and direct detection system without in-line optical amplification and chromatic dispersion compensation. The impacts of variable power loading as well as electrical and optical components on the transmission performance of the demonstrated transceivers are experimentally explored in detail. In addition, numerical simulations also show that variable power loading is an extremely effective means of escalating system performance to its maximum potential.

  5. Evaluation of a composite Gel-Alanine phantom on an end-to-end test to treat multiple brain metastases by a single isocenter VMAT technique.

    PubMed

    Pavoni, Juliana Fernandes; Neves-Junior, Wellington Furtado Pimenta; da Silveira, Matheus Antonio; Haddad, Cecília Maria Kalil; Baffa, Oswaldo

    2017-09-01

    This work aims to evaluate the application of a cylindrical phantom made of dosimetric gel containing alanine pellets distributed inside the gel volume during an end-to-end test of a single isocenter VMAT for simultaneous treatment of multiple brain metastases. The evaluation is based on the comparison of the results obtained with the composite phantom with the treatment planning system (TPS) dose distribution validated by using the clinical conventional quality control with point and planar dose measurements. A cylindrical MAGIC-f gel phantom containing alanine dosimeters (composite phantom) was used to design the VMAT plan in the treatment planning system (TPS). The alanine dosimeters were pellets with radius of 2.5 mm and height of 3 mm, and played the role of brain metastasis inside the gel cylinder, which simulated the cerebral tissue. Five of the alanine dosimeters were selected to simulate five lesions; five planning target volumes (PTVs) were created including the dosimeters and irradiated with different doses. Conventional quality assurance (QA) was performed on the TPS plan and on the composite phantom; a phantom containing only gel (Gel 1 phantom) was also irradiated. One day after irradiation, magnetic resonance images were acquired for both phantoms on a 3T scanner. An electron spin resonance spectrometer was used to evaluate alanine doses. Calibration curves were constructed for the alanine and the gel dosimeters. All the gel only measurement was repeated (Gel 2 phantom) in order to confirm the previous gel measurement. The VMAT treatment plan was approved by the conventional QA. The doses measured by alanine dosimeters on the composite gel phantom agreed to the TPS on average within 3.3%. The alanine dose for each lesion was used to calibrate the gel dosimeter measurements of the concerned PTV. Both gel dose volume histograms (DVH) achieved for each PTV were in agreement with the expected TPS DVH, except for a small discrepancy observed for the Gel 2

  6. The Hurricane-Flood-Landslide Continuum: An Integrated, End-to-end Forecast and Warning System for Mountainous Islands in the Tropics

    NASA Astrophysics Data System (ADS)

    Golden, J.; Updike, R. G.; Verdin, J. P.; Larsen, M. C.; Negri, A. J.; McGinley, J. A.

    2004-12-01

    In the 10 days of 21-30 September 1998, Hurricane Georges left a trail of destruction in the Caribbean region and U.S. Gulf Coast. Subsequently, in the same year, Hurricane Mitch caused widespread destruction and loss of life in four Central American nations, and in December,1999 a tropical disturbance impacted the north coast of Venezuela causing hundreds of deaths and several million dollars of property loss. More recently, an off-season disturbance in the Central Caribbean dumped nearly 250 mm rainfall over Hispaniola during the 24-hr period on May 23, 2004. Resultant flash floods and debris flows in the Dominican Republic and Haiti killed at least 1400 people. In each instance, the tropical system served as the catalyst for major flooding and landslides at landfall. Our goal is to develop and transfer an end-to-end warning system for a prototype region in the Central Caribbean, specifically the islands of Puerto Rico and Hispaniola, which experience frequent tropical cyclones and other disturbances. The envisioned system would include satellite and surface-based observations to track and nowcast dangerous levels of precipitation, atmospheric and hydrological models to predict short-term runoff and streamflow changes, geological models to warn when and where landslides and debris flows are imminent, and the capability to communicate forecast guidance products via satellite to vital government offices in Puerto Rico, Haiti, and the Dominican Republic. In this paper, we shall present a preliminary proof-of-concept study for the May 21-24, 2004 floods and debris-flows over Hispaniola to show that the envisaged flow of data, models and graphical products can produce the desired warning outputs. The multidisciplinary research and technology transfer effort will require blending the talents of hydrometeorologists, geologists, remote sensing and GIS experts, and social scientists to ensure timely delivery of tailored graphical products to both weather offices and local

  7. Endogenous amino nitrogen collected from pigs with end-to-end ileorectal anastomosis is affected by the method of estimation and altered by dietary fiber.

    PubMed

    Mariscal-Landín, G; Sève, B; Colléaux, Y; Lebreton, Y

    1995-01-01

    Endogenous protein loss at the end of the small intestine was determined in two experiments using 10 pigs surgically prepared with end-to-end ileo-rectal anastomosis to allow total collection of ileal digesta. In the first experiment pigs were fed graded protein levels of 0 (protein-free), 55, 110 or 165 g/kg diet. Optimal durations for the adaptation and collection periods were found to be 4 and 3 d, respectively (combination 4:3), as shown by the higher correlation coefficient (r2 = 0.95) between excreted and ingested nitrogen compared with the other combinations tested (5:2, 5:3, 9:3, 9:5). The estimated amounts of endogenous N and amino acids were less accurate and tended to be smaller (P < 0.20) when obtained by extrapolation to zero nitrogen intake than when measured in pigs fed the protein-free diet. The endogenous protein was rich in proline, glutamic acid, glycine, aspartic acid, serine and threonine. In comparison to other amino acid patterns, this composition suggested a low bacterial contamination of the digesta. In the second experiment three levels of dietary fiber from wheat straw, corn cobs and wood cellulose were studied in pigs fed protein-free diets. Between 17 and 34 g crude fiber/kg diet, fiber increased the endogenous losses of nitrogen and amino acids per kilogram of dry matter intake (P < 0.05), but the excretion reached a plateau at higher dietary fiber concentration (102 g/kg). In contrast, glucosamine and galactosamine excretion increased continuously and linearly (P < 0.05) with fiber intake. We conclude that endogenous amino acid loss may be considered constant at usual and high levels of the fibrous mixture under study.

  8. Automated segmentation of 3D anatomical structures on CT images by using a deep convolutional network based on end-to-end learning approach

    NASA Astrophysics Data System (ADS)

    Zhou, Xiangrong; Takayama, Ryosuke; Wang, Song; Zhou, Xinxin; Hara, Takeshi; Fujita, Hiroshi

    2017-02-01

    We have proposed an end-to-end learning approach that trained a deep convolutional neural network (CNN) for automatic CT image segmentation, which accomplished a voxel-wised multiple classification to directly map each voxel on 3D CT images to an anatomical label automatically. The novelties of our proposed method were (1) transforming the anatomical structures segmentation on 3D CT images into a majority voting of the results of 2D semantic image segmentation on a number of 2D-slices from different image orientations, and (2) using "convolution" and "deconvolution" networks to achieve the conventional "coarse recognition" and "fine extraction" functions which were integrated into a compact all-in-one deep CNN for CT image segmentation. The advantage comparing to previous works was its capability to accomplish real-time image segmentations on 2D slices of arbitrary CT-scan-range (e.g. body, chest, abdomen) and produced correspondingly-sized output. In this paper, we propose an improvement of our proposed approach by adding an organ localization module to limit CT image range for training and testing deep CNNs. A database consisting of 240 3D CT scans and a human annotated ground truth was used for training (228 cases) and testing (the remaining 12 cases). We applied the improved method to segment pancreas and left kidney regions, respectively. The preliminary results showed that the accuracies of the segmentation results were improved significantly (pancreas was 34% and kidney was 8% increased in Jaccard index from our previous results). The effectiveness and usefulness of proposed improvement for CT image segmentations were confirmed.

  9. Combined fishing and climate forcing in the southern Benguela upwelling ecosystem: an end-to-end modelling approach reveals dampened effects.

    PubMed

    Travers-Trolet, Morgane; Shin, Yunne-Jai; Shannon, Lynne J; Moloney, Coleen L; Field, John G

    2014-01-01

    The effects of climate and fishing on marine ecosystems have usually been studied separately, but their interactions make ecosystem dynamics difficult to understand and predict. Of particular interest to management, the potential synergism or antagonism between fishing pressure and climate forcing is analysed in this paper, using an end-to-end ecosystem model of the southern Benguela ecosystem, built from coupling hydrodynamic, biogeochemical and multispecies fish models (ROMS-N2P2Z2D2-OSMOSE). Scenarios of different intensities of upwelling-favourable wind stress combined with scenarios of fishing top-predator fish were tested. Analyses of isolated drivers show that the bottom-up effect of the climate forcing propagates up the food chain whereas the top-down effect of fishing cascades down to zooplankton in unfavourable environmental conditions but dampens before it reaches phytoplankton. When considering both climate and fishing drivers together, it appears that top-down control dominates the link between top-predator fish and forage fish, whereas interactions between the lower trophic levels are dominated by bottom-up control. The forage fish functional group appears to be a central component of this ecosystem, being the meeting point of two opposite trophic controls. The set of combined scenarios shows that fishing pressure and upwelling-favourable wind stress have mostly dampened effects on fish populations, compared to predictions from the separate effects of the stressors. Dampened effects result in biomass accumulation at the top predator fish level but a depletion of biomass at the forage fish level. This should draw our attention to the evolution of this functional group, which appears as both structurally important in the trophic functioning of the ecosystem, and very sensitive to climate and fishing pressures. In particular, diagnoses considering fishing pressure only might be more optimistic than those that consider combined effects of fishing and

  10. Design of a satellite end-to-end mission performance simulator for imaging spectrometers and its application to the ESA's FLEX/Sentinel-3 tandem mission

    NASA Astrophysics Data System (ADS)

    Vicent, Jorge; Sabater, Neus; Tenjo, Carolina; Acarreta, Juan R.; Manzano, María.; Rivera, Juan P.; Jurado, Pedro; Franco, Raffaella; Alonso, Luis; Moreno, Jose

    2015-09-01

    The performance analysis of a satellite mission requires specific tools that can simulate the behavior of the platform; its payload; and the acquisition of scientific data from synthetic scenes. These software tools, called End-to-End Mission Performance Simulators (E2ES), are promoted by the European Space Agency (ESA) with the goal of consolidating the instrument and mission requirements as well as optimizing the implemented data processing algorithms. Nevertheless, most developed E2ES are designed for a specific satellite mission and can hardly be adapted to other satellite missions. In the frame of ESA's FLEX mission activities, an E2ES is being developed based on a generic architecture for passive optical missions. FLEX E2ES implements a state-of-the-art synthetic scene generator that is coupled with dedicated algorithms that model the platform and instrument characteristics. This work will describe the flexibility of the FLEX E2ES to simulate complex synthetic scenes with a variety of land cover classes, topography and cloud cover that are observed separately by each instrument (FLORIS, OLCI and SLSTR). The implemented algorithms allows modelling the sensor behavior, i.e. the spectral/spatial resampling of the input scene; the geometry of acquisition; the sensor noises and non-uniformity effects (e.g. stray-light, spectral smile and radiometric noise); and the full retrieval scheme up to Level-2 products. It is expected that the design methodology implemented in FLEX E2ES can be used as baseline for other imaging spectrometer missions and will be further expanded towards a generic E2ES software tool.

  11. Assessing the value of seasonal climate forecast information through an end-to-end forecasting framework: Application to U.S. 2012 drought in central Illinois

    NASA Astrophysics Data System (ADS)

    Shafiee-Jood, Majid; Cai, Ximing; Chen, Ligang; Liang, Xin-Zhong; Kumar, Praveen

    2014-08-01

    This study proposes an end-to-end forecasting framework to incorporate operational seasonal climate forecasts to help farmers improve their decisions prior to the crop growth season, which are vulnerable to unanticipated drought conditions. The framework couples a crop growth model with a decision-making model for rainfed agriculture and translates probabilistic seasonal forecasts into more user-related information that can be used to support farmers' decisions on crop type and some market choices (e.g., contracts with ethanol refinery). The regional Climate-Weather Research and Forecasting model (CWRF) driven by two operational general circulation models (GCMs) is used to provide the seasonal forecasts of weather parameters. To better assess the developed framework, CWRF is also driven by observational reanalysis data, which theoretically can be considered as the best seasonal forecast. The proposed framework is applied to the Salt Creek watershed in Illinois that experienced an extreme drought event during 2012 crop growth season. The results show that the forecasts cannot capture the 2012 drought condition in Salt Creek and therefore the suggested decisions can make farmers worse off if the suggestions are adopted. Alternatively, the optimal decisions based on reanalysis-based CWRF forecasts, which can capture the 2012 drought conditions, make farmers better off by suggesting "no-contract" with ethanol refineries. This study suggests that the conventional metric used for ex ante value assessment is not capable of providing meaningful information in the case of extreme drought. Also, it is observed that institutional interventions (e.g., crop insurance) highly influences farmers' decisions and, thereby, the assessment of forecast value.

  12. OpenCyto: an open source infrastructure for scalable, robust, reproducible, and automated, end-to-end flow cytometry data analysis.

    PubMed

    Finak, Greg; Frelinger, Jacob; Jiang, Wenxin; Newell, Evan W; Ramey, John; Davis, Mark M; Kalams, Spyros A; De Rosa, Stephen C; Gottardo, Raphael

    2014-08-01

    Flow cytometry is used increasingly in clinical research for cancer, immunology and vaccines. Technological advances in cytometry instrumentation are increasing the size and dimensionality of data sets, posing a challenge for traditional data management and analysis. Automated analysis methods, despite a general consensus of their importance to the future of the field, have been slow to gain widespread adoption. Here we present OpenCyto, a new BioConductor infrastructure and data analysis framework designed to lower the barrier of entry to automated flow data analysis algorithms by addressing key areas that we believe have held back wider adoption of automated approaches. OpenCyto supports end-to-end data analysis that is robust and reproducible while generating results that are easy to interpret. We have improved the existing, widely used core BioConductor flow cytometry infrastructure by allowing analysis to scale in a memory efficient manner to the large flow data sets that arise in clinical trials, and integrating domain-specific knowledge as part of the pipeline through the hierarchical relationships among cell populations. Pipelines are defined through a text-based csv file, limiting the need to write data-specific code, and are data agnostic to simplify repetitive analysis for core facilities. We demonstrate how to analyze two large cytometry data sets: an intracellular cytokine staining (ICS) data set from a published HIV vaccine trial focused on detecting rare, antigen-specific T-cell populations, where we identify a new subset of CD8 T-cells with a vaccine-regimen specific response that could not be identified through manual analysis, and a CyTOF T-cell phenotyping data set where a large staining panel and many cell populations are a challenge for traditional analysis. The substantial improvements to the core BioConductor flow cytometry packages give OpenCyto the potential for wide adoption. It can rapidly leverage new developments in computational

  13. An End-to-End simulator for the development of atmospheric corrections and temperature - emissivity separation algorithms in the TIR spectral domain

    NASA Astrophysics Data System (ADS)

    Rock, Gilles; Fischer, Kim; Schlerf, Martin; Gerhards, Max; Udelhoven, Thomas

    2017-04-01

    The development and optimization of image processing algorithms requires the availability of datasets depicting every step from earth surface to the sensor's detector. The lack of ground truth data obliges to develop algorithms on simulated data. The simulation of hyperspectral remote sensing data is a useful tool for a variety of tasks such as the design of systems, the understanding of the image formation process, and the development and validation of data processing algorithms. An end-to-end simulator has been set up consisting of a forward simulator, a backward simulator and a validation module. The forward simulator derives radiance datasets based on laboratory sample spectra, applies atmospheric contributions using radiative transfer equations, and simulates the instrument response using configurable sensor models. This is followed by the backward simulation branch, consisting of an atmospheric correction (AC), a temperature and emissivity separation (TES) or a hybrid AC and TES algorithm. An independent validation module allows the comparison between input and output dataset and the benchmarking of different processing algorithms. In this study, hyperspectral thermal infrared scenes of a variety of surfaces have been simulated to analyze existing AC and TES algorithms. The ARTEMISS algorithm was optimized and benchmarked against the original implementations. The errors in TES were found to be related to incorrect water vapor retrieval. The atmospheric characterization could be optimized resulting in increasing accuracies in temperature and emissivity retrieval. Airborne datasets of different spectral resolutions were simulated from terrestrial HyperCam-LW measurements. The simulated airborne radiance spectra were subjected to atmospheric correction and TES and further used for a plant species classification study analyzing effects related to noise and mixed pixels.

  14. Enzymatic reaction modulated gold nanorod end-to-end self-assembly for ultrahigh sensitively colorimetric sensing of cholinesterase and organophosphate pesticides in human blood.

    PubMed

    Lu, Linlin; Xia, Yunsheng

    2015-08-18

    We present herein the first reported self-assembly modulation of gold nanorods (AuNRs) by enzymatic reaction, which is further employed for colorimetric assays of cholinesterase (ChE) and organophosphate pesticides (OPs) in human blood. ChE catalyzes its substrate (acetylthiocholine) and produces thiocholine and acetate acid. The resulting thiols then react with the tips of the AuNRs by S-Au conjunction and prevent subsequent cysteine-induced AuNR end-to-end (EE) self-assembly. Correspondingly, the AuNR surface plasmon resonance is regulated, which results in a distinctly ratiometric signal output. Under optimal conditions, the linear range is 0.042 to 8.4 μU/mL, and the detection limit is as low as 0.018 μU/mL. As ChE is incubated with OPs, the enzymatic activity is inhibited. So, the cysteine-induced assembly is observed again. On the basis of this principle, OPs can be well determined ranging from 0.12 to 40 pM with a 0.039 pM detection limit. To our knowledge, the present quasi pU/mL level sensitivity for ChE and the quasi femtomolar level sensitivity for OPs are at least 500 and 7000 times lower than those of previous colorimetric methods, respectively. The ultrahigh sensitivity results from (1) the rational choice of anisotropic AuNRs as building blocks and reporters and (2) the specific structure of the enzymatic thiocholine. Because of ultrahigh sensitivity, serum samples are allowed to be extremely diluted in the assay. Accordingly, various nonspecific interactions, even from glutathione/cysteine, are well avoided. So, both ChE and OPs in human blood can be directly assayed without any prepurification, indicating the simplicity and practical promise of the proposed method.

  15. Combined Fishing and Climate Forcing in the Southern Benguela Upwelling Ecosystem: An End-to-End Modelling Approach Reveals Dampened Effects

    PubMed Central

    Travers-Trolet, Morgane; Shin, Yunne-Jai; Shannon, Lynne J.; Moloney, Coleen L.; Field, John G.

    2014-01-01

    The effects of climate and fishing on marine ecosystems have usually been studied separately, but their interactions make ecosystem dynamics difficult to understand and predict. Of particular interest to management, the potential synergism or antagonism between fishing pressure and climate forcing is analysed in this paper, using an end-to-end ecosystem model of the southern Benguela ecosystem, built from coupling hydrodynamic, biogeochemical and multispecies fish models (ROMS-N2P2Z2D2-OSMOSE). Scenarios of different intensities of upwelling-favourable wind stress combined with scenarios of fishing top-predator fish were tested. Analyses of isolated drivers show that the bottom-up effect of the climate forcing propagates up the food chain whereas the top-down effect of fishing cascades down to zooplankton in unfavourable environmental conditions but dampens before it reaches phytoplankton. When considering both climate and fishing drivers together, it appears that top-down control dominates the link between top-predator fish and forage fish, whereas interactions between the lower trophic levels are dominated by bottom-up control. The forage fish functional group appears to be a central component of this ecosystem, being the meeting point of two opposite trophic controls. The set of combined scenarios shows that fishing pressure and upwelling-favourable wind stress have mostly dampened effects on fish populations, compared to predictions from the separate effects of the stressors. Dampened effects result in biomass accumulation at the top predator fish level but a depletion of biomass at the forage fish level. This should draw our attention to the evolution of this functional group, which appears as both structurally important in the trophic functioning of the ecosystem, and very sensitive to climate and fishing pressures. In particular, diagnoses considering fishing pressure only might be more optimistic than those that consider combined effects of fishing and

  16. SU-E-T-360: End-To-End Dosimetric Testing of a Versa HD Linear Accelerator with the Agility Head Modeled in Pinnacle3

    SciTech Connect

    Saenz, D; Narayanasamy, G; Cruz, W; Papanikolaou, N; Stathakis, S

    2015-06-15

    Purpose: The Versa HD incorporates a variety of upgrades, primarily including the Agility head. The distinct dosimetric properties of the head from its predecessors combined with flattening-filter-free (FFF) beams require a new investigation of modeling in planning systems and verification of modeling accuracy. Methods: A model was created in Pinnacle{sup 3} v9.8 with commissioned beam data. Leaf transmission was modeled as <0.5% with maximum leaf speed of 3 cm/s. Photon spectra were tuned for FFF beams, for which profiles were modeled with arbitrary profiles rather than with cones. For verification, a variety of plans with varied parameters were devised, and point dose measurements were compared to calculated values. A phantom of several plastic water and Styrofoam slabs was scanned and imported into Pinnacle{sup 3}. Beams of different field sizes, SSD, wedges, and gantry angles were created. All available photon energies (6 MV, 10 MV, 18 MV, 6 FFF, 10 FFF) as well four clinical electron energies (6, 9, 12, and 15 MeV) were investigated. The plans were verified at a calculation point (8 cm deep for photons, variable for electrons) by measurement with a PTW Semiflex ionization chamber. In addition, IMRT testing was performed with three standard plans (step and shoot IMRT, small and large field VMAT plans). The plans were delivered on the Delta4 IMRT QA phantom (ScandiDos, Uppsala, Sweden). Results: Homogeneous point dose measurement agreed within 2% for all photon and electron beams. Open field photon measurements along the central axis at 100 cm SSD passed within 1%. Gamma passing rates were >99.5% for all plans with a 3%/3mm tolerance criteria. The IMRT QA results for the first 23 patients yielded gamma passing rates of 97.4±2.3%. Conclusion: The end-to-end testing ensured confidence in the ability of Pinnacle{sup 3} to model photon and electron beams with the Agility head.

  17. SU-E-T-19: A New End-To-End Test Method for ExacTrac for Radiation and Plan Isocenter Congruence

    SciTech Connect

    Lee, S; Nguyen, N; Liu, F; Huang, Y; Jung, J; Pyakuryal, A; Jang, S

    2014-06-01

    Purpose: To combine and integrate quality assurance (QA) of target localization and radiation isocenter End to End (E2E) test of BrainLAB ExacTrac system, a new QA approach was devised using anthropomorphic head and neck phantom. This test insures the target localization as well as radiation isocenter congruence which is one step ahead the current ExacTrac QA procedures. Methods: The head and neck phantom typically used for CyberKnife E2E test was irradiated to the sphere target that was visible in CT-sim images. The CT-sim was performed using 1 mm thickness slice with helical scanning technique. The size of the sphere was 3-cm diameter and contoured as a target volume using iPlan V.4.5.2. A conformal arc plan was generated using MLC-based with 7 fields, and five of them were include couch rotations. The prescription dose was 5 Gy and 95% coverage to the target volume. For the irradiation, two Gafchromic films were perpendicularly inserted into the cube that hold sphere inside. The linac used for the irradiation was TrueBeam STx equipped with HD120 MLC. In order to use ExacTrac, infra-red head–array was used to correlate orthogonal X-ray images. Results: Using orthogonal X-rays of ExacTrac the phantom was positioned. For each field, phantom was check again with X-rays and re-positioned if necessary. After each setup using ExacTrac, the target was irradiated. The films were analyzed to determine the deviation of the radiation isocenter in all three dimensions: superior-inferior, left-right and anterior-posterior. The total combining error was found to be 0.76 mm ± 0.05 mm which was within sub-millimeter accuracy. Conclusion: Until now, E2E test for ExacTrac was separately implemented to test image localization and radiation isocenter. This new method can be used for periodic QA procedures.

  18. SU-E-J-25: End-To-End (E2E) Testing On TomoHDA System Using a Real Pig Head for Intracranial Radiosurgery

    SciTech Connect

    Corradini, N; Leick, M; Bonetti, M; Negretti, L

    2015-06-15

    Purpose: To determine the MVCT imaging uncertainty on the TomoHDA system for intracranial radiosurgery treatments. To determine the end-to-end (E2E) overall accuracy of the TomoHDA system for intracranial radiosurgery. Methods: A pig head was obtained from the butcher, cut coronally through the brain, and preserved in formaldehyde. The base of the head was fixed to a positioning plate allowing precise movement, i.e. translation and rotation, in all 6 axes. A repeatability test was performed on the pig head to determine uncertainty in the image bone registration algorithm. Furthermore, the test studied images with MVCT slice thicknesses of 1 and 3 mm in unison with differing scan lengths. A sensitivity test was performed to determine the registration algorithm’s ability to find the absolute position of known translations/rotations of the pig head. The algorithm’s ability to determine absolute position was compared against that of manual operators, i.e. a radiation therapist and radiation oncologist. Finally, E2E tests for intracranial radiosurgery were performed by measuring the delivered dose distributions within the pig head using Gafchromic films. Results: The repeatability test uncertainty was lowest for the MVCTs of 1-mm slice thickness, which measured less than 0.10 mm and 0.12 deg for all axes. For the sensitivity tests, the bone registration algorithm performed better than human eyes and a maximum difference of 0.3 mm and 0.4 deg was observed for the axes. E2E test results in absolute position difference measured 0.03 ± 0.21 mm in x-axis and 0.28 ± 0.18 mm in y-axis. A maximum difference of 0.32 and 0.66 mm was observed in x and y, respectively. The average peak dose difference between measured and calculated dose was 2.7 cGy or 0.4%. Conclusion: Our tests using a pig head phantom estimate the TomoHDA system to have a submillimeter overall accuracy for intracranial radiosurgery.

  19. TU-E-BRB-11: End-To-End Positioning Quality Assurance for Image-Guided Radiosurgery of Multiple Targets Using a Single-Isocenter.

    PubMed

    Popple, R; Fiveash, J; Duan, J; Wu, X; Shen, S; Cardan, R; Brezovich, I

    2012-06-01

    Using a single isocenter significantly reduces delivery times in radiosurgery involving multiple targets. However, because not every target can be placed at isocenter with this type of treatment, a conventional Winston-Lutz test cannot be used. We describe a novel Winston-Lutz like mulitarget test (MTT) for verifying accurate positioning. A target phantom, comprised of an acrylic plate with recesses for three 3/4″ spheres was constructed and a high-resolution (0.5×0.5×0.8 mm) CT scan obtained with PTFE spheres placed in the recesses. The scan was imported into a commercial treatment planning system and multiple beams were prepared, having their isocenter at the centroid of the arrangement of spheres. Every beam incorporated three MLC-defined rectangular apertures that circumscribed the spheres. Custom software selected setup parameters (table, gantry and collimator angle, MLC openings) such that the spheres were centered as precisely as possible within their respective MLC fields, considering the discrete width of collimator leaves. The phantom, with the PTFE replaced by steel spheres, was placed on the treatment couch and imaged using stereoscopic x-ray beams. A 6 degree-of-freedom robotic couch applied translations and rotations to reproduce the CT position. A MV EPID rendered images of the spheres within their respective apertures, allowing identification of sphere and aperture centers. Any error upstream would manifest itself as inaccurate centering of a sphere. Eight beams with table angle 0 and two beams each with table angles 49.7, 89.8, 272.3, and 310.1 were selected. The maximum calculated distance between any sphere and the respective aperture center was 0.07 mm. The median difference measured from the MV images ranged from 0.1 mm to 1.4 mm with a median of 0.8 mm. The MTT is a practical end-to-end test for quality assurance of the entire positioning process in multitarget radiosurgery, from CT scanning to beam delivery. © 2012 American Association

  20. Planning for Mars Sample Return: Results from the MEPAG Mars Sample Return End-to-End International Science Analysis Group (E2E-iSAG)

    NASA Astrophysics Data System (ADS)

    McLennan, S. M.; Sephton, M.; Mepag E2E-Isag

    2011-12-01

    The National Research Council 2011 Planetary Decadal Survey (2013-2022) placed beginning a Mars sample return campaign (MSR) as the top priority for large Flagship missions in the coming decade. Recent developments in NASA-ESA collaborations and Decadal Survey recommendations indicate MSR likely will be an international effort. A joint ESA-NASA 2018 rover (combining the previously proposed ExoMars and MAX-C missions), designed, in part, to collect and cache samples, would thus represent the first of a 3-mission MSR campaign. The End-to-End International Science Analysis Group (E2E-iSAG) was chartered by MEPAG in August 2010 to develop and prioritize MSR science objectives and investigate implications of these objectives for defining the highest priority sample types, landing site selection criteria (and identification of reference landing sites to support engineering planning), requirements for in situ characterization on Mars to support sample selection, and priorities/strategies for returned sample analyses to determine sample sizes and numbers that would meet the objectives. MEPAG approved the E2E-iSAG report in June 2011. Science objectives, summarized in priority order, are: (1) critically assess any evidence for past life or its chemical precursors, and place constraints on past habitability and potential for preservation of signs of life, (2) quantitatively constrain age, context and processes of accretion, early differentiation and magmatic and magnetic history, (3) reconstruct history of surface and near-surface processes involving water, (4) constrain magnitude, nature, timing, and origin of past climate change, (5) assess potential environmental hazards to future human exploration, (6) assess history and significance of surface modifying processes, (7) constrain origin and evolution of the Martian atmosphere, (8) evaluate potential critical resources for future human explorers. All returned samples also would be fully evaluated for extant life as a

  1. Micro-ARES, an electric-field sensor for ExoMars 2016: Electric fields modelling, sensitivity evaluations and end-to-end tests.

    NASA Astrophysics Data System (ADS)

    Déprez, Grégoire; Montmessin, Franck; Witasse, Olivier; Lapauw, Laurent; Vivat, Francis; Abbaki, Sadok; Granier, Philippe; Moirin, David; Trautner, Roland; Hassen-Khodja, Rafik; d'Almeida, Éric; Chardenal, Laurent; Berthelier, Jean-Jacques; Esposito, Francesca; Debei, Stefano; Rafkin, Scott; Barth, Erika

    2014-05-01

    Earth and transposed to the Martian atmospheric parameters. Knowing the expected electric fields and simulating them, the next step in order to evaluate the performance of the instrument is to determine its sensitivity by modelling the response of the instrument. The last step is to confront the model of the instrument, and the expected results for a given signal with the effective outputs of the electric board with the same signal as an input. To achieve this end-to-end test, we use a signal generator followed by an electrical circuit reproducing the electrode behaviour in the Martian environment, in order to inject a realistic electric signal in the processing board and finally compare the produced formatted data with the expected ones.

  2. Experimental demonstration of record high 19.125 Gb/s real-time end-to-end dual-band optical OFDM transmission over 25 km SMF in a simple EML-based IMDD system.

    PubMed

    Giddings, R P; Hugues-Salas, E; Tang, J M

    2012-08-27

    Record high 19.125 Gb/s real-time end-to-end dual-band optical OFDM (OOFDM) transmission is experimentally demonstrated, for the first time, in a simple electro-absorption modulated laser (EML)-based 25 km standard SMF system using intensity modulation and direct detection (IMDD). Adaptively modulated baseband (0-2GHz) and passband (6.125 ± 2GHz) OFDM RF sub-bands, supporting line rates of 10 Gb/s and 9.125 Gb/s respectively, are independently generated and detected with FPGA-based DSP clocked at only 100 MHz and DACs/ADCs operating at sampling speeds as low as 4GS/s. The two OFDM sub-bands are electrically frequency-division-multiplexed (FDM) for intensity modulation of a single optical carrier by an EML. To maximize and balance the signal transmission performance of each sub-band, on-line adaptive features and on-line performance monitoring is fully exploited to optimize key OOFDM transceiver and system parameters, which includes subcarrier characteristics within each individual OFDM sub-band, total and relative sub-band power as well as EML operating conditions. The achieved 19.125 Gb/s over 25 km SMF OOFDM transmission system has an optical power budget of 13.5 dB, and shows almost identical bit error rate (BER) performances for both the baseband and passband signals. In addition, experimental investigations also indicate that the maximum achievable transmission capacity of the present system is mainly determined by the EML frequency chirp-enhanced chromatic dispersion effect, and the passband BER performance is not affected by the two sub-band-induced intermixing effect, which, however, gives a 1.2dB optical power penalty to the baseband signal transmission.

  3. SU-E-T-109: Development of An End-To-End Test for the Varian TrueBeamtm with a Novel Multiple-Dosimetric Modality H and N Phantom

    SciTech Connect

    Zakjevskii, V; Knill, C; Rakowski, J; Snyder, M

    2014-06-01

    Purpose: To develop a comprehensive end-to-end test for Varian's TrueBeam linear accelerator for head and neck IMRT using a custom phantom designed to utilize multiple dosimetry devices. Methods: The initial end-to-end test and custom H and N phantom were designed to yield maximum information in anatomical regions significant to H and N plans with respect to: i) geometric accuracy, ii) dosimetric accuracy, and iii) treatment reproducibility. The phantom was designed in collaboration with Integrated Medical Technologies. A CT image was taken with a 1mm slice thickness. The CT was imported into Varian's Eclipse treatment planning system, where OARs and the PTV were contoured. A clinical template was used to create an eight field static gantry angle IMRT plan. After optimization, dose was calculated using the Analytic Anisotropic Algorithm with inhomogeneity correction. Plans were delivered with a TrueBeam equipped with a high definition MLC. Preliminary end-to-end results were measured using film and ion chambers. Ion chamber dose measurements were compared to the TPS. Films were analyzed with FilmQAPro using composite gamma index. Results: Film analysis for the initial end-to-end plan with a geometrically simple PTV showed average gamma pass rates >99% with a passing criterion of 3% / 3mm. Film analysis of a plan with a more realistic, ie. complex, PTV yielded pass rates >99% in clinically important regions containing the PTV, spinal cord and parotid glands. Ion chamber measurements were on average within 1.21% of calculated dose for both plans. Conclusion: trials have demonstrated that our end-to-end testing methods provide baseline values for the dosimetric and geometric accuracy of Varian's TrueBeam system.

  4. Treatment of a partially thrombosed giant aneurysm of the vertebral artery by aneurysm trapping and direct vertebral artery-posterior inferior cerebellar artery end-to-end anastomosis: technical case report.

    PubMed

    Benes, Ludwig; Kappus, Christoph; Sure, Ulrich; Bertalanffy, Helmut

    2006-07-01

    The purpose of this article is to focus for the first time on the operative management of a direct vertebral artery (VA)-posterior inferior cerebellar artery (PICA) end-to-end anastomosis in a partially thrombosed giant VA-PICA-complex aneurysm and to underline its usefulness as an additional treatment option. The operative technique of a direct VA-PICA end-to-end anatomosis is described in detail. The VA was entering the large aneurysm sack. Distally, the PICA originated from the aneurysm sack-VA-complex. The donor and recipient vessel were cut close to the aneurysm. Whereas the VA was cut in a straight manner, the PICA was cut at an oblique 45-degree angle to enlarge the vascular end diameter. Vessel ends were flushed with heparinized saline and sutured. The thrombotic material inside the aneurysm sack was removed and the distal VA clipped, leaving the anterior spinal artery and brainstem perforators free. The patient regained consciousness without additional morbidity. Magnetic resonance imaging scans revealed a completely decompressed brainstem without infarction. The postoperative angiograms demonstrated a good filling of the anastomosed PICA. Despite the caliber mistmatch of these two vessels the direct VA-PICA end-to-end anastomosis provides an accurate alternative in addition to other anastomoses and bypass techniques, when donor and recipient vessels are suitable and medullary perforators do not have to be disrupted.

  5. Effects of knee flexion angle and loading conditions on the end-to-end distance of the posterior cruciate ligament: a comparison of the roles of the anterolateral and posteromedial bundles.

    PubMed

    Wang, Joon Ho; Kato, Yuki; Ingham, Sheila J M; Maeyama, Akira; Linde-Rosen, Monica; Smolinski, Patrick; Fu, Freddie H; Harner, Christopher

    2014-12-01

    It is commonly accepted that the anterolateral (AL) bundle of the posterior cruciate ligament (PCL) is tight in flexion and that the posteromedial (PM) bundle is tight in extension. However, a recent in vivo study showed that both bundles were tight in extension. To investigate the effects of knee flexion angle, rotational torque, and anterior/posterior translational force on the end-to-end distance between the femoral and tibial insertion sites of each bundle of the PCL. Descriptive laboratory study. Cadaveric knees (10 specimens) were mounted on a robotic system, and the relative positional data between the femur and tibia were acquired during passive flexion/extension, with an applied 5-N·m rotational torque and an applied 89-N translational force. The bony surface and PCL insertion data were acquired with a 3-dimensional scanner after gross dissection and were superimposed onto the positional data. The end-to-end distance between the 2 PCL insertion sites of the femur and tibia was measured. The end-to-end distance increased from full extension to 90° for both the AL (9.2 ± 1.8 mm; from 30.0 to 39.2 mm) and PM bundles (5.8 ± 2.2 mm; from 32.0 to 37.7 mm). With an internal rotational torque, the end-to-end distance of the PM bundle increased significantly (P < .05) at 0°, 30°, and 60° of knee flexion. Under a posterior translational force at 90° of knee flexion, the length of both bundles increased to their longest measurements (AM bundle: 40.6 ± 4.2 mm; PM bundle: 38.4 ± 3.8 mm). The end-to-end distance of the AL and PM bundles of the PCL increased in flexion, and this pattern was maintained during tests with posterior translational force. The PM bundle was more affected by the rotational torque than was the AL bundle. Both bundles of the PCL may serve a greater functional role in flexion than in extension. The PM bundle might be more important for the control of rotation than the AL bundle. Posterior translation at 90° of knee flexion could be the

  6. 21 CFR 172.876 - Castor oil.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 21 Food and Drugs 3 2011-04-01 2011-04-01 false Castor oil. 172.876 Section 172.876 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED) FOOD FOR HUMAN... Multipurpose Additives § 172.876 Castor oil. The food additive castor oil may be safely used in accordance with...

  7. 21 CFR 172.876 - Castor oil.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... 21 Food and Drugs 3 2014-04-01 2014-04-01 false Castor oil. 172.876 Section 172.876 Food and Drugs... PERMITTED FOR DIRECT ADDITION TO FOOD FOR HUMAN CONSUMPTION Multipurpose Additives § 172.876 Castor oil. The food additive castor oil may be safely used in accordance with the following conditions: (a) The...

  8. 21 CFR 172.876 - Castor oil.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... 21 Food and Drugs 3 2013-04-01 2013-04-01 false Castor oil. 172.876 Section 172.876 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED) FOOD FOR HUMAN... Multipurpose Additives § 172.876 Castor oil. The food additive castor oil may be safely used in accordance with...

  9. 21 CFR 172.876 - Castor oil.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... 21 Food and Drugs 3 2012-04-01 2012-04-01 false Castor oil. 172.876 Section 172.876 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED) FOOD FOR HUMAN... Multipurpose Additives § 172.876 Castor oil. The food additive castor oil may be safely used in accordance with...

  10. 21 CFR 172.876 - Castor oil.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 21 Food and Drugs 3 2010-04-01 2009-04-01 true Castor oil. 172.876 Section 172.876 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED) FOOD FOR HUMAN... Multipurpose Additives § 172.876 Castor oil. The food additive castor oil may be safely used in accordance...

  11. Toxin content of commercial castor cultivars

    USDA-ARS?s Scientific Manuscript database

    The castor plant Ricinus communis L. is the source of castor oil which has numerous uses and is a key chemical feedstock for an array of products from polymers to cosmetics. Although castor was once widely grown throughout the world, the presence of the toxic protein ricin has deterred the re-introd...

  12. Demonstration of end-to-end cloud-DSL with a PON-based fronthaul supporting 5.76-Gb/s throughput with 48 eCDMA-encoded 1024-QAM discrete multi-tone signals.

    PubMed

    Fang, Liming; Zhou, Lei; Liu, Xiang; Zhang, Xiaofeng; Sui, Meng; Effenberger, Frank; Zhou, Jun

    2015-05-18

    We experimentally demonstrate an end-to-end ultra-broadband cloud-DSL network using passive optical network (PON) based fronthaul with electronic code-division-multiple-access (eCDMA) encoding and decoding. Forty-eight signals that are compliant with the very-high-bit-rate digital subscriber line 2 (VDSL2) standard are transmitted with a record throughput of 5.76 Gb/s over a hybrid link consisting of a 20-km standard single-mode fiber and a 100-m twisted pair.

  13. Polynomial End-to-End Communication

    DTIC Science & Technology

    1989-08-01

    the "fail-stop model" in [AM88]. As mentioned in the introduction, our motivation here is to deal with networks frequently changing topology. In such...eventually connected edges among nodes. By Q4, there is no cut between the receiver and sender in G’. It is known, from graph theory , that if is there is no...Computing, Toronto, Ontario, Canada, pages 131-148, ACM SIGACT and SIGOPS, ACM, 1988. 22 [AGH89] Baruch Awerbuch, Oded Goldreich, and Amir Herzberg . A

  14. Distributed and End-to-End Testing

    DTIC Science & Technology

    2009-12-01

    TACCOMM Tactical Communications TAS MK-23 Target Acquisition System Mark-23 TPX -42 The Interrogator Set - 42 TR Trouble Report VA Virginia...Automatic Correlation (SGS/AC), Target Acquisition System Mark-23 (TAS MK-23), Interrogator Set ( TPX -42), Integrated Automated Detection and Tracking

  15. A comparative study of red and blue light-emitting diodes and low-level laser in regeneration of the transected sciatic nerve after an end to end neurorrhaphy in rabbits.

    PubMed

    Takhtfooladi, Mohammad Ashrafzadeh; Sharifi, Davood

    2015-12-01

    This study aimed at evaluating the effects of red and blue light-emitting diodes (LED) and low-level laser (LLL) on the regeneration of the transected sciatic nerve after an end-to-end neurorrhaphy in rabbits. Forty healthy mature male New Zealand rabbits were randomly assigned into four experimental groups: control, LLL (680 nm), red LED (650 nm), and blue LED (450 nm). All animals underwent the right sciatic nerve neurotmesis injury under general anesthesia and end-to-end anastomosis. The phototherapy was initiated on the first postoperative day and lasted for 14 consecutive days at the same time of the day. On the 30th day post-surgery, the animals whose sciatic nerves were harvested for histopathological analysis were euthanized. The nerves were analyzed and quantified the following findings: Schwann cells, large myelinic axons, and neurons. In the LLL group, as compared to other groups, an increase in the number of all analyzed aspects was observed with significance level (P < 0.05). This finding suggests that postoperative LLL irradiation was able to accelerate and potentialize the peripheral nerve regeneration process in rabbits within 14 days of irradiation.

  16. Effect of glyphosate on the castor plant Ricinus communis L

    USDA-ARS?s Scientific Manuscript database

    The presence of a toxin in the castor seed has been a major impediment to the reintroduction of castor as a domestic crop in the US. Because dehiscent forms of castor are invasive and castor seed remains viable for years under the right conditions, there is concern that if castor were widely cultiva...

  17. Castor oil: Biosynthesis and Uses

    USDA-ARS?s Scientific Manuscript database

    Castor oil is unique among commodity oils in its fatty acid composition, which consists of 90% ricinoleate, (12-hydroxy-octadec-cis 9-enoate). The mid-chain hydroxyl group imparts physical and chemical properties that make it useful in many industrial applications. Among its uses are lithium grease,...

  18. Syntheses, structures, and magnetic properties of three one-dimensional end-to-end azide/cyanate-bridged copper(II) compounds exhibiting ferromagnetic interaction: new type of solid state isomerism.

    PubMed

    Sasmal, Sujit; Sarkar, Sohini; Aliaga-Alcalde, Núria; Mohanta, Sasankasekhar

    2011-06-20

    The work in this paper presents the syntheses, structures, and magnetic properties of three end-to-end (EE) azide/cyanate-bridged copper(II) compounds [Cu(II)L(1)(μ(1,3)-NCO)](n)·2nH(2)O (1), [Cu(II)L(1)(μ(1,3)-N(3))](n)·2nH(2)O (2), and [Cu(II)L(2)(μ(1,3)-N(3))](n) (3), where the ligands used to achieve these species, HL(1) and HL(2), are the tridentate Schiff base ligands obtained from [1 + 1] condensations of salicylaldehyde with 4-(2-aminoethyl)-morpholine and 3-methoxy salicylaldehyde with 1-(2-aminoethyl)-piperidine, respectively. Compounds 1 and 2 crystallize in the monoclinic P2(1)/c space group, while compound 3 crystallizes in the orthorhombic Pbca space group. The metal center in 1-3 is in all cases pentacoordinated. Three coordination positions of the metal center in 1, 2, or 3 are satisfied by the phenoxo oxygen atom, imine nitrogen atom, and morpholine (for 1 and 2) or piperidine (for 3) nitrogen atom of one deprotonated ligand, [L(1)](-) or [L(2)](-). The remaining two coordination positions are satisfied by two nitrogen atoms of two end-to-end bridging azide ligands for 2 and 3 and one nitrogen atom and one oxygen atom of two end-to-end bridging cyanate ligands for 1. The coordination geometry of the metal ion is distorted square pyramidal in which one EE azide/cyanate occupies the apical position. Variable-temperature (2-300 K) magnetic susceptibilities of 1-3 have been measured under magnetic fields of 0.05 (from 2 to 30 K) and 1.0 T (from 30 to 300 K). The simulation reveals a ferromagnetic interaction in all three compounds with J values of +0.19 ± 0.01, +0.79 ± 0.01, and +1.25 ± 0.007 cm(-1) for 1, 2, and 3, respectively. Compound 1 is the sole example of a ferromagnetically coupled EE cyanate-bridged 1-D copper(II) system. In addition, a rare example of supramolecular isomerism and a nice example of magnetic isomerism have been observed and most interestingly a new type of solid state isomerism has emerged as a result of the comparison

  19. Unmanned Aircraft Systems Detect and Avoid System: End-to-End Verification and Validation Simulation Study of Minimum Operations Performance Standards for Integrating Unmanned Aircraft into the National Airspace System

    NASA Technical Reports Server (NTRS)

    Ghatas, Rania W.; Jack, Devin P.; Tsakpinis, Dimitrios; Sturdy, James L.; Vincent, Michael J.; Hoffler, Keith D.; Myer, Robert R.; DeHaven, Anna M.

    2017-01-01

    As Unmanned Aircraft Systems (UAS) make their way to mainstream aviation operations within the National Airspace System (NAS), research efforts are underway to develop a safe and effective environment for their integration into the NAS. Detect and Avoid (DAA) systems are required to account for the lack of "eyes in the sky" due to having no human on-board the aircraft. The technique, results, and lessons learned from a detailed End-to-End Verification and Validation (E2-V2) simulation study of a DAA system representative of RTCA SC-228's proposed Phase I DAA Minimum Operational Performance Standards (MOPS), based on specific test vectors and encounter cases, will be presented in this paper.

  20. Efficacy and safety of a NiTi CAR 27 compression ring for end-to-end anastomosis compared with conventional staplers: A real-world analysis in Chinese colorectal cancer patients.

    PubMed

    Lu, Zhenhai; Peng, Jianhong; Li, Cong; Wang, Fulong; Jiang, Wu; Fan, Wenhua; Lin, Junzhong; Wu, Xiaojun; Wan, Desen; Pan, Zhizhong

    2016-05-01

    This study aimed to evaluate the safety and efficacy of a new nickel-titanium shape memory alloy compression anastomosis ring, NiTi CAR 27, in constructing an anastomosis for colorectal cancer resection compared with conventional staples. In total, 234 consecutive patients diagnosed with colorectal cancer receiving sigmoidectomy and anterior resection for end-to-end anastomosis from May 2010 to June 2012 were retrospectively analyzed. The postoperative clinical parameters, postoperative complications and 3-year overall survival in 77 patients using a NiTi CAR 27 compression ring (CAR group) and 157 patients with conventional circular staplers (STA group) were compared. There were no statistically significant differences between the patients in the two groups in terms of general demographics and tumor features. A clinically apparent anastomotic leak occurred in 2 patients (2.6%) in the CAR group and in 5 patients (3.2%) in the STA group (p=0.804). These eight patients received a temporary diverting ileostomy. One patient (1.3%) in the CAR group was diagnosed with anastomotic stricture through an electronic colonoscopy after 3 months postoperatively. The incidence of postoperative intestinal obstruction was comparable between the two groups (p=0.192). With a median follow-up duration of 39.6 months, the 3-year overall survival rate was 83.1% in the CAR group and 89.0% in the STA group (p=0.152). NiTi CAR 27 is safe and effective for colorectal end-to-end anastomosis. Its use is equivalent to that of the conventional circular staplers. This study suggests that NiTi CAR 27 may be a beneficial alternative in colorectal anastomosis in Chinese colorectal cancer patients.

  1. Safety and efficacy of the NiTi Shape Memory Compression Anastomosis Ring (CAR/ColonRing) for end-to-end compression anastomosis in anterior resection or low anterior resection.

    PubMed

    Kang, Jeonghyun; Park, Min Geun; Hur, Hyuk; Min, Byung Soh; Lee, Kang Young; Kim, Nam Kyu

    2013-04-01

    Compression anastomoses may represent an improvement over traditional hand-sewn or stapled techniques. This prospective exploratory study aimed to assess the efficacy and complication rates in patients undergoing anterior resection (AR) or low anterior resection (LAR) anastomosed with a novel end-to-end compression anastomosis ring, the ColonRing. In all, 20 patients (13 male) undergoing AR or LAR were enrolled to be anastomosed using the NiTi Shape Memory End-to-End Compression Anastomosis Ring (NiTi Medical Technologies Ltd, Netanya, Israel). Demographic, intraoperative, and postoperative data were collected. Patients underwent AR (11/20) or LAR using laparoscopy (75%), robotic (10%) surgery, or an open laparotomy (15%) approach, with a median anastomotic level of 14.5 cm (range, 4-25 cm). Defunctioning loop ileostomies were formed in 6 patients for low anastomoses. Surgeons rated the ColonRing device as either easy or very easy to use. One patient developed an anastomotic leakage in the early postoperative period; there were no late postoperative complications. Mean time to passage of first flatus and commencement of oral fluids was 2.5 days and 3.2 days, respectively. Average hospital stay was 12.6 days (range, 8-23 days). Finally, the device was expelled on average 15.3 days postoperatively without difficulty. This is the first study reporting results in a significant number of LAR patients and the first reported experience from South Korea; it shows that the compression technique is surgically feasible, easy to use, and without significant complication rates. A large randomized controlled trial is warranted to investigate the benefits of the ColonRing over traditional stapling techniques.

  2. Castor Oil-Based Biodegradable Polyesters.

    PubMed

    Kunduru, Konda Reddy; Basu, Arijit; Haim Zada, Moran; Domb, Abraham J

    2015-09-14

    This Review compiles the synthesis, physical properties, and biomedical applications for the polyesters based on castor oil and ricinoleic acid. Castor oil has been known for its medicinal value since ancient times. It contains ∼90% ricinoleic acid, which enables direct chemical transformation into polyesters without interference of other fatty acids. The presence of ricinoleic acid (hydroxyl containing fatty acid) enables synthesis of various polyester/anhydrides. In addition, castor oil contains a cis-double bond that can be hydrogenated, oxidized, halogenated, and polymerized. Castor oil is obtained pure in large quantities from natural sources; it is safe and biocompatible.

  3. SU-F-P-37: Implementation of An End-To-End QA Test of the Radiation Therapy Imaging, Planning and Delivery Process to Identify and Correct Possible Sources of Deviation

    SciTech Connect

    Salinas Aranda, F; Suarez, V; Arbiser, S; Sansogne, R

    2016-06-15

    Purpose: To implement an end-to-end QA test of the radiation therapy imaging, planning and delivery process, aimed to assess the dosimetric agreement accuracy between planned and delivered treatment, in order to identify and correct possible sources of deviation. To establish an internal standard for machine commissioning acceptance. Methods: A test involving all steps of the radiation therapy: imaging, planning and delivery process was designed. The test includes analysis of point dose and planar dose distributions agreement between TPS calculated and measured dose. An ad hoc 16 cm diameter PMMA phantom was constructed with one central and four peripheral bores that can accommodate calibrated electron density inserts. Using Varian Eclipse 10.0 and Elekta XiO 4.50 planning systems, IMRT, RapidArc and 3DCRT with hard and dynamic wedges plans were planned on the phantom and tested. An Exradin A1SL chamber is used with a Keithley 35617EBS electrometer for point dose measurements in the phantom. 2D dose distributions were acquired using MapCheck and Varian aS1000 EPID.Gamma analysis was performed for evaluation of 2D dose distribution agreement using MapCheck software and Varian Portal Dosimetry Application.Varian high energy Clinacs Trilogy, 2100C/CD, 2000CR and low energy 6X/EX where tested.TPS-CT# vs. electron density table were checked for CT-scanners used. Results: Calculated point doses were accurate to 0.127% SD: 0.93%, 0.507% SD: 0.82%, 0.246% SD: 1.39% and 0.012% SD: 0.01% for LoX-3DCRT, HiX-3DCRT, IMRT and RapidArc plans respectively. Planar doses pass gamma 3% 3mm in all cases and 2% 2mm for VMAT plans. Conclusion: Implementation of a simple and reliable quality assurance tool was accomplished. The end-to-end proved efficient, showing excellent agreement between planned and delivered dose evidencing strong consistency of the whole process from imaging through planning to delivery. This test can be used as a first step in beam model acceptance for clinical

  4. DEVELOPING A SAFE SOURCE OF CASTOR OIL

    USDA-ARS?s Scientific Manuscript database

    Castor bean (Ricinus communis L.) is an important oilseed crop with significant industrial value. However, the production of castor oil is hampered by the presence of the toxin ricin and hyper-allergenic 2S albumins in its seed. We are thus investigating the possibility of developing a safe source...

  5. Synthesis of novel castor oil phosphonates

    USDA-ARS?s Scientific Manuscript database

    Castor oil has served as a versatile hydroxy fatty acid (HFA); its principle component, ricinoleic acid, can be isolated from castor oil and has been modified extensively for a number of applications. Additionally, phosphonates and their corresponding phosphonic acids are a functional moiety that ha...

  6. Biodegradable foam plastics based on castor oil.

    PubMed

    Wang, Hong Juan; Rong, Min Zhi; Zhang, Ming Qiu; Hu, Jing; Chen, Hui Wen; Czigány, Tibor

    2008-02-01

    In this work, a simple but effective approach was proposed for preparing biodegradable plastic foams with a high content of castor oil. First of all, castor oil reacted with maleic anhydride to produce maleated castor oil (MACO) without the aid of any catalyst. Then plastic foams were synthesized through free radical initiated copolymerization between MACO and diluent monomer styrene. With changes in MACO/St ratio and species of curing initiator, mechanical properties of MACO foams can be easily adjusted. In this way, biofoams with comparable compressive stress at 25% strain as commercial polyurethane (PU) foams were prepared, while the content of castor oil can be as high as 61 wt %. The soil burial tests further proved that the castor oil based foams kept the biodegradability of renewable resources despite the fact that some petrol-based components were introduced.

  7. Single-fraction spine SBRT end-to-end testing on TomoTherapy, Vero, TrueBeam, and CyberKnife treatment platforms using a novel anthropomorphic phantom.

    PubMed

    Gallo, John J; Kaufman, Isaac; Powell, Rachel; Pandya, Shalini; Somnay, Archana; Bossenberger, Todd; Ramirez, Ezequiel; Reynolds, Robert; Solberg, Timothy; Burmeister, Jay

    2015-01-08

    Spine SBRT involves the delivery of very high doses of radiation to targets adjacent to the spinal cord and is most commonly delivered in a single fraction. Highly conformal planning and accurate delivery of such plans is imperative for successful treatment without catastrophic adverse effects. End-to-end testing is an important practice for evaluating the entire treatment process from simulation through treatment delivery. We performed end-to-end testing for a set of representative spine targets planned and delivered using four different treatment planning systems (TPSs) and delivery systems to evaluate the various capabilities of each. An anthropomorphic E2E SBRT phantom was simulated and treated on each system to evaluate agreement between measured and calculated doses. The phantom accepts ion chambers in the thoracic region and radiochromic film in the lumbar region. Four representative targets were developed within each region (thoracic and lumbar) to represent different presentations of spinal metastases and planned according to RTOG 0631 constraints. Plans were created using the TomoTherapy TPS for delivery using the Hi·Art system, the iPlan TPS for delivery using the Vero system, the Eclipse TPS for delivery using the TrueBeam system in both flattened and flattening filter free (FFF), and the MultiPlan TPS for delivery using the CyberKnife system. Delivered doses were measured using a 0.007 cm3 ion chamber in the thoracic region and EBT3 GAFCHROMIC film in the lumbar region. Films were scanned and analyzed using an Epson Expression 10000XL flatbed scanner in conjunction with FilmQAPro2013. All treatment platforms met all dose constraints required by RTOG 0631. Ion chamber measurements in the thoracic targets delivered an overall average difference of 1.5%. Specifically, measurements agreed with the TPS to within 2.2%, 3.2%, 1.4%, 3.1%, and 3.0% for all three measureable cases on TomoTherapy, Vero, TrueBeam (FFF), TrueBeam (flattened), and Cyber

  8. Direct evidence for the occurrence of simple microtubule organizing centers (MTOCs) and for end-to-end joining of microtubules with modified ends in the reticulopodia of Allogromia laticollaris (foraminifera).

    PubMed

    Golz, R; Hauser, M

    1991-11-29

    The partial disassembly of filopodial microtubules into helical filaments, respectively the reassembly from paracrystals of exactly registered tubulin helices was recognized as a special adaptation to the highly dynamic reticulopodial movements in the foraminifer Allogromia laticollaris [4, 16]. Electron microscopic investigations of partially extracted or completely lysed filopodial strands revealed some additional adaptations to the local tubulin demands within the widely spread, rapidly changing reticulopodial networks. As probable foci for the initiation of additional microtubules from pools of tubulin during the netto growth phase, small (40 × 60 nm) cube-shaped, electron dense bodies were discovered which could act as mobile MTOCs at very distant sites from the metabolically active cell body within the reticulopodial system. The microtubule ends, proximally as well as distally to the filopodial origins, are characterized by a closed appearance because of a terminal cap which is continuous with the microtubule wall, whereas some other microtubules appear as the usual open cylinders. It is suggested that the closed ends, which occur either freely or decorated by the cube-shaped, electron dense bodies, may play a role in the de novo initiation of microtubules as well as in the transformation to the helical state or vice versa. Points of discontinuity within single longitudinally cut microtubules can be observed directly by electron microscopy and correspond to end-to-end contacts of closed MT-terminations. These junctions probably represent initial contact points of an annealing process between adjacent shorter MT-segments. Microtubule annealing occurs rapidly in vitro [14, 15] and could explain the elongation rates observed during filopodial outgrowth. Copyright © 1991 Gustav Fischer Verlag · Stuttgart · Jena · New York. Published by Elsevier GmbH.. All rights reserved.

  9. The Dosimetric Importance of Six Degree of Freedom Couch End to End Quality Assurance for SRS/SBRT Treatments when Comparing Intensity Modulated Radiation Therapy to Volumetric Modulated Arc Therapy

    NASA Astrophysics Data System (ADS)

    Ulizio, Vincent Michael

    With the advancement of technology there is an increasing ability for lesions to be treated with higher radiation doses each fraction. This also allows for low fractionated treatments. Because the patient is receiving a higher dose of radiation per fraction and because of the fast dose falloff in these targets there must be extreme accuracy in the delivery. The 6 DOF couch allows for extra rotational corrections and for a more accurate set-up. The movement of the couch needs to be verified to be accurate and because of this, end to end quality assurance tests for the couch have been made. After the set-up is known to be accurate then different treatment techniques can be studied. SBRT of the Spine has a very fast dose falloff near the spinal cord and was typically treated with IMRT. Treatment plans generated using this technique tend to have streaks of low dose radiation, so VMAT is being studied to determine if this treatment technique can reduce the low dose radiation volume as well as improve OAR sparing. For the 6 DOF couch QA, graph paper is placed on the anterior and right lateral sides of the VisionRT OSMS Cube Phantom. Each rotational shift is then applied individually, with a 3 degree shift in the positive and negative directions for pitch and roll. A mark is drawn on the paper to record each shift. A CBCT is then taken of the Cube and known shifts are applied and then an additional CBCT is taken to return the Cube to isocenter. The original IMRT plans for SBRT of the Spine are evaluated and then a plan is made utilizing VMAT. These plans are then compared for low dose radiation, OAR sparing, and conformity. If the original IMRT plan is determined to be an inferior treatment to what is acceptable, then this will be re-planned and compared to the VMAT plan. The 6 DOF couch QA tests have proven to be accurate and reproducible. The average deviations in the 3 degree and -3 degree pitch and roll directions were 0.197, 0.068, 0.091, and 0.110 degrees

  10. Study and Implementation of the End-to-End Data Pipeline for the Virtis Imaging Spectrometer Onbaord Venus Express: "From Science Operations Planning to Data Archiving and Higher Lever Processing"

    NASA Astrophysics Data System (ADS)

    Cardesín Moinelo, Alejandro

    2010-04-01

    This PhD Thesis describes the activities performed during the Research Program undertaken for two years at the Istituto Nazionale di AstroFisica in Rome, Italy, as active member of the VIRTIS Technical and Scientific Team, and one additional year at the European Space Astronomy Center in Madrid, Spain, as member of the Mars Express Science Ground Segment. This document will show a study of all sections of the Science Ground Segment of the Venus Express mission, from the planning of the scientific operations, to the generation, calibration and archiving of the science data, including the production of valuable high level products. We will present and discuss here the end-to-end diagram of the ground segment from the technical and scientific point of view, in order to describe the overall flow of information: from the original scientific requests of the principal investigator and interdisciplinary teams, up to the spacecraft, and down again for the analysis of the measurements and interpretation of the scientific results. These scientific results drive to new and more elaborated scientific requests, which are used as feedback to the planning cycle, closing the circle. Special attention is given here to describe the implementation and development of the data pipeline for the VIRTIS instrument onboard Venus Express. During the research program, both the raw data generation pipeline and the data calibration pipeline were developed and automated in order to produce the final raw and calibrated data products from the input telemetry of the instrument. The final raw and calibrated products presented in this work are currently being used by the VIRTIS Science team for data analysis and are distributed to the whole scientific community via the Planetary Science Archive. More than 20,000 raw data files and 10,000 calibrated products have already been generated after almost 4 years of mission. In the final part of the Thesis, we will also present some high level data

  11. 75 FR 44917 - Castor Oil, Ethoxylated, Dioleate; Tolerance Exemption

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-07-30

    ...-0232; FRL-8835-3] Castor Oil, Ethoxylated, Dioleate; Tolerance Exemption AGENCY: Environmental... requirement of a tolerance for residues of castor oil, ethoxylated, dioleate; when used as an inert ingredient... permissible level for residues of castor oil, ethoxylated, dioleate on food or feed commodities. DATES:...

  12. 21 CFR 178.3280 - Castor oil, hydrogenated.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... 21 Food and Drugs 3 2012-04-01 2012-04-01 false Castor oil, hydrogenated. 178.3280 Section 178.3280 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED... SANITIZERS Certain Adjuvants and Production Aids § 178.3280 Castor oil, hydrogenated. Hydrogenated castor oil...

  13. 21 CFR 178.3280 - Castor oil, hydrogenated.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... 21 Food and Drugs 3 2014-04-01 2014-04-01 false Castor oil, hydrogenated. 178.3280 Section 178.3280 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED... § 178.3280 Castor oil, hydrogenated. Hydrogenated castor oil may be safely used in the manufacture of...

  14. 21 CFR 178.3280 - Castor oil, hydrogenated.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... 21 Food and Drugs 3 2013-04-01 2013-04-01 false Castor oil, hydrogenated. 178.3280 Section 178.3280 Food and Drugs FOOD AND DRUG ADMINISTRATION, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED... SANITIZERS Certain Adjuvants and Production Aids § 178.3280 Castor oil, hydrogenated. Hydrogenated castor oil...

  15. Rapid development of a castor cultivar with increased oil content

    USDA-ARS?s Scientific Manuscript database

    Castor seed oil contains 90% ricinoleic acid which has a wide range of industrial applications. Improvement in oil content would be of great benefit to castor growers and oil processers. Two cycles of phenotypic recurrent selection were conducted through screening for high oil content castor seeds u...

  16. Reducing the toxicity of castor seed meal through processing treatments

    USDA-ARS?s Scientific Manuscript database

    The castor plant produces a seed that is high in oil content and composed of approximately 90% ricinoleate. Due to the numerous uses of castor oil and ricinoleate, the oil is in high demand. However, the presence of a protein toxin in the seed meal is a key concern about processing the castor seed t...

  17. Enzymatic incorporation of caffeoyl into castor oil to prepare the novel castor oil-based caffeoyl structured lipids.

    PubMed

    Sun, Shangde; Wang, Ping; Zhu, Sha

    2017-05-10

    In this work, a novel castor oil-based caffeoyl structured lipids was successfully prepared by the enzymatic transesterification using castor oil (CO) as caffeoyl acceptor. During the structured lipids preparation, two competitive reactions, the hydrolysis of CO to form hydrophilic caffeoyl glycerols (CG)+dicaffeoyl glycerols (DCG) and the transesterification of CO with ethyl caffeate (EC) to form lipophilic caffeoyl mono- and di-acylglycerols (CMAG and CDAG), were found. Reaction progress was monitored using HPLC-ESI-MS and HPLC-UV. The effects of by-product ethanol removal and reaction variables on the transesterification and reaction selectivity were evaluated. Results showed that, the activation energies for the transesterification and for the selective formations of CMAG+CDAG and CG+DCG were 57.60kJ/mol, 58.86kJ/mol, and 60.53kJ/mol, respectively. Under the optimal reaction conditions (enzyme load 23%, 90°C, 1:3 molar ratio of EC to CO, and 46.5h), EC conversion and the yield of CMAG+CDAG were 93.68±2.52% and 78.11±1.35%, respectively. Copyright © 2017 Elsevier B.V. All rights reserved.

  18. Phenylpropanoid esters of lesquerella and castor oil

    USDA-ARS?s Scientific Manuscript database

    Lesquerella (LO) and castor oil (CO) were esterified at the secondary hydroxyl groups of their 14-hydroxyeicos-cis-11-enoic fatty acids and 12-hydroxyoctadec-cis-9-enoic fatty acids, respectively, with 4-acetoxy-3-methoxycinnamic acid (acetoxyferulic acid). The unconventional esterifications were co...

  19. Tough Blends of Polylactide and Castor Oil

    SciTech Connect

    Robertson, Megan L.; Paxton, Jessica M.; Hillmyer, Marc A.

    2012-10-10

    Poly(l-lactide) (PLLA) is a renewable resource polymer derived from plant sugars with several commercial applications. Broader implementation of the material is limited due to its inherent brittleness. We show that the addition of 5 wt % castor oil to PLLA significantly enhances the overall tensile toughness with minimal reductions in the modulus and no plasticization of the PLLA matrix. In addition, we used poly(ricinoleic acid)-PLLA diblock copolymers, synthesized entirely from renewable resources, as compatibilizers for the PLLA/castor oil blends. Ricinoleic acid, the majority fatty acid comprising castor oil, was polymerized through a lipase-catalyzed condensation reaction. The resulting polymers contained a hydroxyl end-group that was subsequently used to initiate the ring-opening polymerization of L-lactide. The binary PLLA/castor oil blend exhibited a tensile toughness seven times greater than neat PLLA. The addition of block copolymer allowed for control over the morphology of the blends, and even further improvement in the tensile toughness was realized - an order of magnitude larger than that of neat PLLA.

  20. Tough blends of polylactide and castor oil.

    PubMed

    Robertson, Megan L; Paxton, Jessica M; Hillmyer, Marc A

    2011-09-01

    Poly(l-lactide) (PLLA) is a renewable resource polymer derived from plant sugars with several commercial applications. Broader implementation of the material is limited due to its inherent brittleness. We show that the addition of 5 wt % castor oil to PLLA significantly enhances the overall tensile toughness with minimal reductions in the modulus and no plasticization of the PLLA matrix. In addition, we used poly(ricinoleic acid)-PLLA diblock copolymers, synthesized entirely from renewable resources, as compatibilizers for the PLLA/castor oil blends. Ricinoleic acid, the majority fatty acid comprising castor oil, was polymerized through a lipase-catalyzed condensation reaction. The resulting polymers contained a hydroxyl end-group that was subsequently used to initiate the ring-opening polymerization of l-lactide. The binary PLLA/castor oil blend exhibited a tensile toughness seven times greater than neat PLLA. The addition of block copolymer allowed for control over the morphology of the blends, and even further improvement in the tensile toughness was realized-an order of magnitude larger than that of neat PLLA.

  1. Effect of castor oil emulsion eyedrops on tear film composition and stability.

    PubMed

    Maïssa, Cécile; Guillon, Michel; Simmons, Peter; Vehige, Joseph

    2010-04-01

    An emulsion eyedrop containing castor oil has been shown to modify the tear film lipid layer and increase tear film stability. The primary objectives of this investigation were to measure the prevalence of castor oil in the tear fluid over time and quantify the effects on the lipid layer. A secondary objective was to quantify the initial effects on ocular symptomatology. The investigation was an open label pilot study on 5 normal and 10 dry eye subjects. A single eyedrop (Castor oil emulsion, Allergan) was instilled in each eye; the tear film appearance and composition were monitored for 4h via in vivo visualisation using the Tearscope and post in vivo tear samples analysis by HPLC. Combined results for both normal and dry eye subjects showed that castor oil was detected up to 4h after a single eyedrop instillation and associated with an increase in the level of tear film lipid. The relative amount of various lipid families was also changed. An increase in tear lipid layer thickness was significant up to one hour post-instillation for the symptomatic sub-population. The changes in tear film characteristics were associated with significantly lower symptoms up to four hours post-instillation for the symptomatic sub-population. This pilot investigation showed that castor oil eyedrops achieved a residence time of at least four hours post-instillation, producing a more stable tear film and an associated significant decrease in ocular symptoms over the entire follow-up period for the symptomatic subjects. 2009 British Contact Lens Association. Published by Elsevier Ltd. All rights reserved.

  2. Development of Castor-Oil-Resistant Polyurethane Sonar Encapsulants.

    DTIC Science & Technology

    2014-09-26

    7D-S? 576 DEVELOPMENT OF CASTOR- OIL -RESISTANT POLYURETHANE SONAR 1/1 ENCRPSULANTS(U) NRVAL RESEARCH LAB MASHINGTON DC UNCLSSIFED T MA3URICE ET AL. 30...ACCESSION NO. Washington, DC 23062 64503N I S0219AS (59)0584 i1 TITLE (Include Security Classification) Development of Castor- Oil -Resistant...identify by block number) FIELD GROUP SUB-GROUP Castor oil -- Sonar transducer encapsulants 11 07 - Polyurethane\\ K , 19. ABSTRACT (Continue on reverse if

  3. Thiokol Corporation's Castor 120 solid propulsion booster program

    NASA Astrophysics Data System (ADS)

    Hilden, J.

    1992-07-01

    The Castor 120 program aimed at developing a solid propulsion booster for expandable launch vehicles is discussed. The booster is based on a light weight epoxy case, HTPB propellant, and flexible bearing nozzle. The Castor 120 will be capable of satisfying multiple launch vehicle functions for use as a strap-on, a first- or a second stage.

  4. Evaluation of Genetic Diversity of Castor Bean for Biodiesel Utilization

    USDA-ARS?s Scientific Manuscript database

    Castor bean (Ricinus communis L., 2n=20) is a cross-pollinated diploid species belonging to the family Euphorbiaceae instead of the Leguminosae. It is a native of Africa but may have originated in India. Castor bean plants grow as annual or perennial, depending on geographical locations, climate a...

  5. Novel alpha-hydroxy phosphonic acids via castor oil

    USDA-ARS?s Scientific Manuscript database

    Hydroxy fatty acids (HFAs) have found a number of uses in today’s market, with uses ranging from materials to pharmaceuticals. Castor oil has served as a versatile HFA; its principle component, ricinoleic acid, can be isolated from castor oil and has been modified extensively for a number of applica...

  6. Improved method for extraction of castor seed for toxin determination

    USDA-ARS?s Scientific Manuscript database

    The effort to identify castor seeds with low ricin content is considered to be a key to increasing cultivation of the castor plant for industrial applications. The procedure used to obtain soluble protein from the seed is a limiting factor for screening large numbers of seeds. Usually, the seed is...

  7. Risk of ricin from commercial castor production in North America

    USDA-ARS?s Scientific Manuscript database

    Commercial production of castor, as a source of highly valuable hydroxyl fatty acids, has been limited by both the real and perceived risks of commercial castor production in North America. Crop commodity groups, regulatory governmental agencies, and much of the general public may have reservations ...

  8. End-to-End Service Oriented Architectures (SOA) Security Project

    DTIC Science & Technology

    2012-02-01

    architecture for SOA. • A novel service invocation control mechanism for SOA using dynamic taint analysis (TA). • A trust broker (TB) system that maintains...architectures (SOAs) because the SOAs stress on machine-to-machine interactions, while most of the IT security mechanisms are based on human-to- machine...listing services worldwide. It is a standard mechanism for registering or publishing and discovering Web services. The services published into the UDDI

  9. Ordered End-to-End Multicast for Distributed Multimedia Systems

    DTIC Science & Technology

    2006-01-01

    Ordering of messages compensates for the lack of a global system state and the effects of asynchrony, unpredictable network delay, and disparities in...separate logical propagation graph or global clock synchronization, and ordering is dis- tributed across nodes on the delivery paths between sources...set on route by PN, deciding on a globally valid num- ber, and multicasting the message to the receiver set with a final and binding sequence number

  10. End-to-end experiment management in HPC

    SciTech Connect

    Bent, John M; Kroiss, Ryan R; Torrez, Alfred; Wingate, Meghan

    2010-01-01

    Experiment management in any domain is challenging. There is a perpetual feedback loop cycling through planning, execution, measurement, and analysis. The lifetime of a particular experiment can be limited to a single cycle although many require myriad more cycles before definite results can be obtained. Within each cycle, a large number of subexperiments may be executed in order to measure the effects of one or more independent variables. Experiment management in high performance computing (HPC) follows this general pattern but also has three unique characteristics. One, computational science applications running on large supercomputers must deal with frequent platform failures which can interrupt, perturb, or terminate running experiments. Two, these applications typically integrate in parallel using MPI as their communication medium. Three, there is typically a scheduling system (e.g. Condor, Moab, SGE, etc.) acting as a gate-keeper for the HPC resources. In this paper, we introduce LANL Experiment Management (LEM), an experimental management framework simplifying all four phases of experiment management. LEM simplifies experiment planning by allowing the user to describe their experimental goals without having to fully construct the individual parameters for each task. To simplify execution, LEM dispatches the subexperiments itself thereby freeing the user from remembering the often arcane methods for interacting with the various scheduling systems. LEM provides transducers for experiments that automatically measure and record important information about each subexperiment; these transducers can easily be extended to collect additional measurements specific to each experiment. Finally, experiment analysis is simplified by providing a general database visualization framework that allows users to quickly and easily interact with their measured data.

  11. Kepler Mission: End-to-End System Demonstration

    NASA Technical Reports Server (NTRS)

    Borucki, William; Koch, D.; Dunham, E.; Jenkins, J.; Witteborn, F.; Updike, T.; DeVincenzi, Donald L. (Technical Monitor)

    2000-01-01

    A test facility has been constructed to demonstrate the capability of differential ensemble photometry to detect transits of Earth-size planets orbiting solar-like stars. The main objective is to determine the effects of various noise sources on the capability of a CCD photometer to maintain a system relative precision of 1 x $10^(-5)$ for mv = 12 stars in the presence of system-induced noise sources. The facility includes a simulated star field, fast optics to simulate the telescope, a thinned back-illuminated CCD similar to those to be used on the spacecraft and computers to perform the onboard control, data processing and extraction. The test structure is thermally and mechanically isolated so that each source of noise can be introduced in a controlled fashion and evaluated for its contribution to the total noise budget. The effects of pointing errors or a changing thermal environment are imposed by piezo-electric devices. Transits are injected by heating small wires crossing apertures in the star plate. Signals as small as those from terrestrial-size transits of solar-like stars are introduced to demonstrate that such planets can be detected under realistic noise conditions. Examples of imposing several noise sources and the resulting detectabilities are presented. These show that a differential ensemble photometric approach CCD photometer can readily detect signals associated with Earth-size transits.

  12. Analysis of End-to-End Performance of LAN Systems

    DTIC Science & Technology

    1990-03-01

    thesis research. We show the results with respect to LAN utilization, request delay, complete transfer, delivery time , and incomplete transfer. These...request delay, LAN utilization and delivery time will be measured for the purpose of the performance analysis of LANs. Since the analytical approach based...one or two servers (as Tables 7-12 and Figures 26-31) " AVG, STD DEV delivery time for transaction class 1 with one or two servers: from PC to Server

  13. End-to-End Performance Management for Large Distributed Storage

    SciTech Connect

    Almadena Chtchelkanova

    2012-03-18

    Storage systems for large distributed clusters of computer servers are themselves large and distributed. Their complexity and scale make it hard to ensure that applications using them get good, predictable performance. At the same time, shared access to the system from multiple applications, users, and internal system activities leads to a need for predictable performance. This research investigates mechanisms for improving storage system performance in large distributed storage systems through mechanisms that integrate the performance aspects of the path that I/O operations take through the system, from the application interface on the compute server, through the network, to the storate servers. The research focuses on five parts of the I/O path in a distributed storage system: I/O scheduling at the storage server, storage server cache management, client-to-server network flow control, client-to-server connection management, and client cache management.

  14. Going End to End to Deliver High-Speed Data

    NASA Technical Reports Server (NTRS)

    2005-01-01

    By the end of the 1990s, the optical fiber "backbone" of the telecommunication and data-communication networks had evolved from megabits-per-second transmission rates to gigabits-per-second transmission rates. Despite this boom in bandwidth, however, users at the end nodes were still not being reached on a consistent basis. (An end node is any device that does not behave like a router or a managed hub or switch. Examples of end node objects are computers, printers, serial interface processor phones, and unmanaged hubs and switches.) The primary reason that prevents bandwidth from reaching the end nodes is the complex local network topology that exists between the optical backbone and the end nodes. This complex network topology consists of several layers of routing and switch equipment which introduce potential congestion points and network latency. By breaking down the complex network topology, a true optical connection can be achieved. Access Optical Networks, Inc., is making this connection a reality with guidance from NASA s nondestructive evaluation experts.

  15. On Estimating End-to-End Network Path Properties

    NASA Technical Reports Server (NTRS)

    Allman, Mark; Paxson, Vern

    1999-01-01

    The more information about current network conditions available to a transport protocol, the more efficiently it can use the network to transfer its data. In networks such as the Internet, the transport protocol must often form its own estimates of network properties based on measurements per-formed by the connection endpoints. We consider two basic transport estimation problems: determining the setting of the retransmission timer (RTO) for are reliable protocol, and estimating the bandwidth available to a connection as it begins. We look at both of these problems in the context of TCP, using a large TCP measurement set [Pax97b] for trace-driven simulations. For RTO estimation, we evaluate a number of different algorithms, finding that the performance of the estimators is dominated by their minimum values, and to a lesser extent, the timer granularity, while being virtually unaffected by how often round-trip time measurements are made or the settings of the parameters in the exponentially-weighted moving average estimators commonly used. For bandwidth estimation, we explore techniques previously sketched in the literature [Hoe96, AD98] and find that in practice they perform less well than anticipated. We then develop a receiver-side algorithm that performs significantly better.

  16. Biosynthesis of castor oil studied by the regiospecific analysis of castor triacylglycerols by ESI-MS

    USDA-ARS?s Scientific Manuscript database

    HPLC fractions of diricinoleoyl-acyl-glycerols containing one non-ricinoleoyl chain from castor oil were used to identify the regiospecific location of this non-ricinoleoyl chain on the glycerol backbone using electrospray ionization-MS3 of lithium adducts. The regiospecific ions used were from the ...

  17. Castor oil for induction of labour: a retrospective study.

    PubMed

    Neri, Isabella; Dante, Giulia; Pignatti, Lucrezia; Salvioli, Chiara; Facchinetti, Fabio

    2017-06-15

    The aim of this study is to investigate the safety and efficacy of castor oil to induce labour. A retrospective observational case control study was conducted over five years. Castor oil was proposed to women referred to the Birth Centre (Castor Oil group (COG)). They were compared to women who chose to be followed by the traditional doctor-led unit (control group (CG)). Castor oil was administered in a 60 ml single dose in 200 ml of warm water. Inclusion criteria were gestational age between 40 and 41 weeks plus premature rupture of membranes between 12 and 18 hours or amniotic fluid index ≤4 or Bishop Score of ≤4 or absence of spontaneous labour over 41 + 4 weeks. Pharmacological induction of labour was required for 18 women in the COG (45%) and 36 in the CG (90%) (p < .001). The mode of delivery differed significantly between groups: women assuming castor oil showed a higher incidence of vaginal delivery, whereas the incidence of caesarean section was lower in the COG, but no statistical significance was reached. The use of castor oil is related to a higher probability of labour initiation within 24 hours. Castor oil can be considered a safe non-pharmacological method for labour induction.

  18. Streamlining CASTOR to manage the LHC data torrent

    NASA Astrophysics Data System (ADS)

    Lo Presti, G.; Espinal Curull, X.; Cano, E.; Fiorini, B.; Ieri, A.; Murray, S.; Ponce, S.; Sindrilaru, E.

    2014-06-01

    This contribution describes the evolution of the main CERN storage system, CASTOR, as it manages the bulk data stream of the LHC and other CERN experiments, achieving over 90 PB of stored data by the end of LHC Run 1. This evolution was marked by the introduction of policies to optimize the tape sub-system throughput, going towards a cold storage system where data placement is managed by the experiments' production managers. More efficient tape migrations and recalls have been implemented and deployed where bulk meta-data operations greatly reduce the overhead due to small files. A repack facility is now integrated in the system and it has been enhanced in order to automate the repacking of several tens of petabytes, required in 2014 in order to prepare for the next LHC run. Finally the scheduling system has been evolved to integrate the internal monitoring. To efficiently manage the service a solid monitoring infrastructure is required, able to analyze the logs produced by the different components (about 1 kHz of log messages). A new system has been developed and deployed, which uses a transport messaging layer provided by the CERN-IT Agile Infrastructure and exploits technologies including Hadoop and HBase. This enables efficient data mining by making use of MapReduce techniques, and real-time data aggregation and visualization. The outlook for the future is also presented. Directions and possible evolution will be discussed in view of the restart of data taking activities.

  19. Castor Oil Based Polyurethanes: Synthesis and Characterization

    NASA Astrophysics Data System (ADS)

    Macalino, AD; Salen, VA; Reyes, LQ

    2017-09-01

    In this study, polyurethanes based on castor oil and 1,6-hexamethylene diisocyanate (HMDI) were synthesized with varying weight ratio of the castor oil and HMDI. The formation of urethane linkages was verified through the use of a fourier transform infrared spectroscopy (FTIR). The hydrophilicity of the films was evaluated through the use of a contact angle meter and it was found that the contact angle of all the films were below 90 degrees which confirms their hydrophilicity. The thermal stability of the PU films were studies through the use of a thermal gravimetric analyzer and found that all of the polyurethane films exhibited two weight loss events at elevated temperatures wherein the first weight loss event was observed to occur at 285°C to 384°C while the second weight loss event was observed at around 521°C to 551°C. The hardness, elastic modulus, and tensile elongation of the PU films were determined by using a universal testing machine (UTM) where it was found out that the hardness and the elastic modulus of the film is directly proportional with HMDI loading while the tensile elongation is inversely proportional to it. Lastly, it was known through the swelling studies of the PU films that it does not swell, this is due to the presence of unreacted triglycerides in the material, which prevents water from permeating to the films.

  20. Castor sacs and anal glands of the north american beaver (Castor canadensis): their histology, development, and relationship to scent communication.

    PubMed

    Walro, J M; Svendsen, G E

    1982-05-01

    Both sexes of beavers possess a pair of castor sacs and a pair of anal glands located in paired subcutaneous cavities between the pelvis and the base of the tail. The castor sacs are not glandular in the histological sense, hence references to these structures as preputial glands or castor glands are misnomers. The wall of the castor sacs is plicate and comprised of three distinct zones: an outer layer of vascular connective tissue, a two-to five-cell-thick layer of mitotic epithelial cells, and several densely packed layers of cornified epithelium which grade into more widely separated sheets toward the lumen. Monocultures of a gram-positive facultatively anaerobic bacterium were present in the lumen of all castor sac preparations. Differences in the frequency of castoreum deposition were not attributable to differences in the structure of the castor sacs. The anal glands of beavers are holocrine sebaceous glands. These glands develop more rapidly than the castor sacs. Anal gland tissue from embryos exhibited cellular characteristics associated with the production of sebum. Secretory activity was evident in all preparations. The relationship of castoreum and anal gland secretion to scent communication among beavers is discussed.

  1. 28. Photocopy of engineering drawing. CASTOR IV MODIFICATIONS LAUNCH COMPLEX ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    28. Photocopy of engineering drawing. CASTOR IV MODIFICATIONS LAUNCH COMPLEX 17A: LEVEL 1A PLATFORMS-STRUCTURAL, 1973. - Cape Canaveral Air Station, Launch Complex 17, Facility 28416, East end of Lighthouse Road, Cape Canaveral, Brevard County, FL

  2. 29. Photocopy of engineering drawing. CASTOR IV MODIFICATIONS LAUNCH COMPLEX ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    29. Photocopy of engineering drawing. CASTOR IV MODIFICATIONS LAUNCH COMPLEX 17A: LEVEL 17A PLATFORMS-STRUCTURAL, 1973. - Cape Canaveral Air Station, Launch Complex 17, Facility 28416, East end of Lighthouse Road, Cape Canaveral, Brevard County, FL

  3. Renewable polyethylene mimics derived from castor oil.

    PubMed

    Türünç, Oĝuz; Montero de Espinosa, Lucas; Meier, Michael A R

    2011-09-01

    An increasing number of reports on the syntheses of carbohydrate- and plant oil-based polymers has been published in ongoing efforts to produce plastic materials from renewable resources. Although many of these polymers are biodegradable and this is a desirable property for certain applications, in some cases non-degradable polymers are needed for long-term use purposes. Polyolefins are one of the most important classes of materials that have already taken their places in our daily life. On the other hand, their production relies on fossil resources. Therefore, within this contribution, we discuss synthetic routes toward a number of polyethylene mimics derived from fatty acids via thiol-ene and ADMET polymerization reactions in order to establish more sustainable routes toward this important class of polymers. Two different diene monomers were thus prepared from castor oil derived platform chemicals, their polymerization via the two mentioned routes was optimized and compared to each other, and their thermal properties were investigated.

  4. Castor Bean Organelle Genome Sequencing and Worldwide Genetic Diversity Analysis

    PubMed Central

    Chan, Agnes P.; Williams, Amber L.; Rice, Danny W.; Liu, Xinyue; Melake-Berhan, Admasu; Huot Creasy, Heather; Puiu, Daniela; Rosovitz, M. J.; Khouri, Hoda M.; Beckstrom-Sternberg, Stephen M.; Allan, Gerard J.; Keim, Paul; Ravel, Jacques; Rabinowicz, Pablo D.

    2011-01-01

    Castor bean is an important oil-producing plant in the Euphorbiaceae family. Its high-quality oil contains up to 90% of the unusual fatty acid ricinoleate, which has many industrial and medical applications. Castor bean seeds also contain ricin, a highly toxic Type 2 ribosome-inactivating protein, which has gained relevance in recent years due to biosafety concerns. In order to gain knowledge on global genetic diversity in castor bean and to ultimately help the development of breeding and forensic tools, we carried out an extensive chloroplast sequence diversity analysis. Taking advantage of the recently published genome sequence of castor bean, we assembled the chloroplast and mitochondrion genomes extracting selected reads from the available whole genome shotgun reads. Using the chloroplast reference genome we used the methylation filtration technique to readily obtain draft genome sequences of 7 geographically and genetically diverse castor bean accessions. These sequence data were used to identify single nucleotide polymorphism markers and phylogenetic analysis resulted in the identification of two major clades that were not apparent in previous population genetic studies using genetic markers derived from nuclear DNA. Two distinct sub-clades could be defined within each major clade and large-scale genotyping of castor bean populations worldwide confirmed previously observed low levels of genetic diversity and showed a broad geographic distribution of each sub-clade. PMID:21750729

  5. The Castor 120 (TM) motor: Development and qualification testing results

    NASA Technical Reports Server (NTRS)

    Hilden, Jack G.; Poirer, Beverly M.

    1993-01-01

    This paper discusses Thiokol Corporation's static test results for the development and qualification program of the Castor 120(TM) motor. The demonstration program began with a 25,000-pound motor to demonstrate the new technologies and processes that would be used on the larger Castor 120(TM) motor. The Castor 120(TM) motor was designed to be applicable as a first stage, second stage, or strap-on motor. Static test results from the Castor 25 and two Castor 120(TM) motors are discussed in this paper. The results verified the feasibility of tailoring the propellant grain configuration and nozzle throat diameter to meet various customer requirements. The first and second motors were conditioned successfully at ambient temperature and 28 F, respectively, to demonstrate that the design could handle a wide range of environmental launch conditions. Furthermore, the second Castor 120(TM) motor demonstrated a systems tunnel and forward skirt extension to verify flight-ready stage hardware. It is anticipated that the first flight motor will be ready by the fall of 1994.

  6. CastorDB: a comprehensive knowledge base for Ricinus communis

    PubMed Central

    2011-01-01

    Background Ricinus communis is an industrially important non-edible oil seed crop, native to tropical and subtropical regions of the world. Although, R. communis genome was assembled in 4X draft by JCVI, and is predicted to contain 31,221 proteins, the function of most of the genes remains to be elucidated. A large amount of information of different aspects of the biology of R. communis is available, but most of the data are scattered one not easily accessible. Therefore a comprehensive resource on Castor, Castor DB, is required to facilitate research on this important plant. Findings CastorDB is a specialized and comprehensive database for the oil seed plant R. communis, integrating information from several diverse resources. CastorDB contains information on gene and protein sequences, gene expression and gene ontology annotation of protein sequences obtained from a variety of repositories, as primary data. In addition, computational analysis was used to predict cellular localization, domains, pathways, protein-protein interactions, sumoylation sites and biochemical properties and has been included as derived data. This database has an intuitive user interface that prompts the user to explore various possible information resources available on a given gene or a protein. Conclusion CastorDB provides a user friendly comprehensive resource on castor with particular emphasis on its genome, transcriptome, and proteome and on protein domains, pathways, protein localization, presence of sumoylation sites, expression data and protein interacting partners. PMID:21914200

  7. Castor bean organelle genome sequencing and worldwide genetic diversity analysis.

    PubMed

    Rivarola, Maximo; Foster, Jeffrey T; Chan, Agnes P; Williams, Amber L; Rice, Danny W; Liu, Xinyue; Melake-Berhan, Admasu; Huot Creasy, Heather; Puiu, Daniela; Rosovitz, M J; Khouri, Hoda M; Beckstrom-Sternberg, Stephen M; Allan, Gerard J; Keim, Paul; Ravel, Jacques; Rabinowicz, Pablo D

    2011-01-01

    Castor bean is an important oil-producing plant in the Euphorbiaceae family. Its high-quality oil contains up to 90% of the unusual fatty acid ricinoleate, which has many industrial and medical applications. Castor bean seeds also contain ricin, a highly toxic Type 2 ribosome-inactivating protein, which has gained relevance in recent years due to biosafety concerns. In order to gain knowledge on global genetic diversity in castor bean and to ultimately help the development of breeding and forensic tools, we carried out an extensive chloroplast sequence diversity analysis. Taking advantage of the recently published genome sequence of castor bean, we assembled the chloroplast and mitochondrion genomes extracting selected reads from the available whole genome shotgun reads. Using the chloroplast reference genome we used the methylation filtration technique to readily obtain draft genome sequences of 7 geographically and genetically diverse castor bean accessions. These sequence data were used to identify single nucleotide polymorphism markers and phylogenetic analysis resulted in the identification of two major clades that were not apparent in previous population genetic studies using genetic markers derived from nuclear DNA. Two distinct sub-clades could be defined within each major clade and large-scale genotyping of castor bean populations worldwide confirmed previously observed low levels of genetic diversity and showed a broad geographic distribution of each sub-clade.

  8. CASTOR: Structural dynamics in the MIR station

    NASA Astrophysics Data System (ADS)

    Vialaneix, J. P.; Bousquet, P.; Dancet, Y.; Guay, P.; Mercier, F.

    1996-10-01

    Prediction and reduction of space structure vibrations have become critical issues over the last ten years. Motivating factors include optical systems with high stability requirements, large antennas, sensitive micro gravity payloads, compatibility of flexible appendages with attitude control, and large orbital infrastructure surviving. The experiment "CASTOR" (French acronym for ChAracterisation of STructures in ORbit) is dedicated to the identification of the structural dynamic modes of the MIR station and to the investigation of the dynamic behaviour—in zero g conditions—of a truss mock-up equipped with various passive and active damping technologies. The measured modal parameters of MIR are to be compared with the results of a finite element model analysis. The differences between in-flight and on-ground dynamics of the truss will be thoroughly analysed. This project has been conceived and is managed by CNES. The flight hardware will be delivered by the end of 1995, and the experimental work will be performed by a French cosmonaut within the framework of the CASSIOPEE mission in June 1996. In the first place, a comprehensive overview of CNES activities in the prediction of in-flight structural dynamics, and of similar experimental efforts found in literature will be discussed. Afterwards, the motivations behind the CASTOR experiment will be shown, followed by a full description of its equipment. The results of the ground tests and analyses, and the in-orbit test plan will then be presented extensively. Particular emphasis will be laid on the performances of the active damping systems which will be validated in flight. In conclusion, the potential re-utilisation of the experiment material within the framework of later flights will be highlighted.

  9. CASTOR damping experiment in-flight results

    NASA Astrophysics Data System (ADS)

    Guay, Philippe; Bousquet, Paul W.; Courau, E.; Mercier, F.

    2001-07-01

    Over the past ten years, space structures such as optical systems, large antennae or sensitive microgravity payloads have increased their stability requirements. To answer this need, prediction and reduction of vibration levels have been widely studied. The 'CASTOR' truss experiment (French acronym for ChAracterization of STructures in ORbit) is mainly dedicated to the investigation of the dynamic behaviour -in zero g conditions- of a truss mock-up with various damping technologies. The performances of various dampers have been validated in flight, measurements compared with the predictions and the results analysed for different technological solutions. This project was developed under CNES management, and the final experimental work was performed in orbit by a French cosmonaut on MIR station during the summer of 1999. In the first place, the French space agency's interest in structural dynamics behind the CASTOR experiment will be recalled. The flight hardware will then be fully described and a short clip filmed during the PERSEUS mission in MIR will supported this. The truss test bench will be presented, together with the design drivers for damping the first structural modes below 100Hz. The dampers -which are efficient- are located in some strategic bars in the truss. Their performance has been optimized using an energetic criterion. This paper will then focus on the design of the damping, using either piezoelectric active control or passive energy dissipation in elastomer or fluid devices. Particular emphasis will be laid upon the excellent flight results obtained with all the various damping systems, and on the effect of in-orbit conditions. A comparison between the different technologies here tested will also be made with regards to performances achieved and their suitability for space applications. In conclusion, lessons learned during the development process will be highlighted and further work on spacecraft applications discussed.

  10. Energy requirements of beavers (Castor canadensis) swimming underwater.

    PubMed

    Allers, D; Culik, B M

    1997-01-01

    Energy requirements of beavers (Castor canadensis) swimming voluntarily underwater were investigated in Neumünster Zoo (Germany) in a covered, still-water swim channel with oxygen and carbon dioxide respirometry. During the experiments, all activities of the beavers were monitored and recorded. While at rest within their thermoneutral zone on land (17 degrees C), beavers had a respiratory quotient of 0.95 and a resting metabolic rate of 1.58 W kg-1. When resting in water, energy requirements rose to 2.31 W kg-1. When swimming underwater in the channel, beavers preferred a mean speed of 0.64 m s-1, and their energy requirements rose to 2.64 W kg-1. Cost of transport, however, was minimal at 0.9 m s-1 and amounted to 0.36 J N-1 m-1. Although beavers must compromise form and function to operate on water and on land, their energy requirements while diving amount to only 1.65 times the resting metabolic rate and compare well with those of accomplished swimmers such as aquatic mammals and birds.

  11. Castor Oil: Properties, Uses, and Optimization of Processing Parameters in Commercial Production

    PubMed Central

    Patel, Vinay R.; Dumancas, Gerard G.; Kasi Viswanath, Lakshmi C.; Maples, Randall; Subong, Bryan John J.

    2016-01-01

    Castor oil, produced from castor beans, has long been considered to be of important commercial value primarily for the manufacturing of soaps, lubricants, and coatings, among others. Global castor oil production is concentrated primarily in a small geographic region of Gujarat in Western India. This region is favorable due to its labor-intensive cultivation method and subtropical climate conditions. Entrepreneurs and castor processors in the United States and South America also cultivate castor beans but are faced with the challenge of achieving high castor oil production efficiency, as well as obtaining the desired oil quality. In this manuscript, we provide a detailed analysis of novel processing methods involved in castor oil production. We discuss novel processing methods by explaining specific processing parameters involved in castor oil production. PMID:27656091

  12. Castor Oil: Properties, Uses, and Optimization of Processing Parameters in Commercial Production.

    PubMed

    Patel, Vinay R; Dumancas, Gerard G; Kasi Viswanath, Lakshmi C; Maples, Randall; Subong, Bryan John J

    2016-01-01

    Castor oil, produced from castor beans, has long been considered to be of important commercial value primarily for the manufacturing of soaps, lubricants, and coatings, among others. Global castor oil production is concentrated primarily in a small geographic region of Gujarat in Western India. This region is favorable due to its labor-intensive cultivation method and subtropical climate conditions. Entrepreneurs and castor processors in the United States and South America also cultivate castor beans but are faced with the challenge of achieving high castor oil production efficiency, as well as obtaining the desired oil quality. In this manuscript, we provide a detailed analysis of novel processing methods involved in castor oil production. We discuss novel processing methods by explaining specific processing parameters involved in castor oil production.

  13. WIDER DISSEMINATION OF CASTOR BEAN ALLERGEN—Factors Presaging Increasing Incidence of Disease in California

    PubMed Central

    Small, Willard S.

    1953-01-01

    With the growing, transportation and processing of castor beans in California rapidly in creasing, it is probable that the incidence of allergic disease owing to sensitivity to the castor bean allergen also will increase. PMID:13019607

  14. Castor oil for induction of labour: not harmful, not helpful.

    PubMed

    Boel, Machteld Elisabeth; Lee, Sue Jean; Rijken, Marcus Johannes; Paw, Moo Koo; Pimanpanarak, Mupawjay; Tan, Saw Oo; Singhasivanon, Pratap; Nosten, François; McGready, Rose

    2009-10-01

    Castor oil is one of the most popular drugs for induction of labour in a non-medical setting; however, published data on safety and effectiveness of this compound to induce labour remain sparse. To assess the safety and effectiveness of castor oil for induction of labour in pregnancies with an ultrasound estimated gestational at birth of more than 40 weeks. Data were extracted from hospital-based records of all pregnant women who attended antenatal clinics on the Thai-Burmese border and who were more than 40 weeks pregnant. The effectiveness of castor oil to induce labour was expressed as time to birth and analysed with a Cox proportional hazards regression model. Measures associated with safety were fetal distress, meconium-stained amniotic fluid, tachysystole of the uterus, uterine rupture, abnormal maternal blood pressure during labour, Apgar scores, neonatal resuscitation, stillbirth, post-partum haemorrhage, severe diarrhoea and maternal death. Proportions were compared using Fisher's exact test. Of 612 women with a gestation of more than 40 weeks, 205 received castor oil for induction and 407 did not. The time to birth was not significantly different between the two groups (hazard ratio 0.99 (95% confidence interval: 0.81 to 1.20; n = 509)). Castor oil use was not associated with any harmful effects on the mother or fetus. Castor oil for induction of labour had no effect on time to birth nor were there any harmful effects observed in this large series. Our findings leave no justification for recommending castor oil for this purpose.

  15. Searching for a Safe Source of Castor Oil Production through Metabolic Engineering

    USDA-ARS?s Scientific Manuscript database

    Castor oil contains 90% ricinoleate (12-hydroxy-oleate) which has numerous industrial uses. The production of castor oil is hampered by the presence of the toxin ricin and hyper-allergenic 2S albumins in its seed. We are developing a safe source of castor oil by two approaches: blocking gene expres...

  16. Analysis of acylglycerols containing mono- and dihydroxy fatty acids in castor oil by HPLC and MS

    USDA-ARS?s Scientific Manuscript database

    Ricinoleate, a monohydroxy fatty acid, has many industrial uses such as the manufacture of aviation lubricant, plastic, paint and cosmetics. Ricinoleate occurs as acylglycerols (AG) in castor oil, and about 70% of castor oil is triricinolein. Castor oil is the only commercial source of ricinoleate. ...

  17. Regiospecific Quantification of Triacylglycerols Containing Ricinoleate and Dihydroxy Fatty Acids in Castor Oil by Mass Spectrometry

    USDA-ARS?s Scientific Manuscript database

    Ricinoleate, a monohydroxy fatty acid, has many industrial uses such as the manufacture of aviation lubricant, plastic, paint and cosmetics. Ricinoleate occurs as acylglycerols (AG) in castor oil, and about 70% of castor oil is triricinolein. Castor oil is the only commercial source of ricinoleate. ...

  18. Effects of leupeptin on proteinase and germination of castor beans

    SciTech Connect

    Alpi, A.; Beevers, H.

    1981-10-01

    Leupeptin, tripeptide inhibitor of some proteinases, was shown previously to maintain the stability of several enzymes (isocitrate lyase, fumarase, and catalase) in crude extracts of castor bean endosperm. This reagent is now shown to inhibit the breakdown of water-soluble and crystalloid-storage proteins of the protein bodies isolated from castor beans by the SH-proteinase and it also inhibits the endopeptidase from mung beans. When suitably introduced into the endosperm of dry castor beans it strongly inhibits germination and seedling development. Application of leupeptin to endosperm halves removed from the seed prevents the normal development of enzymes concerned with gluconeogenesis from fat and drastically curtails sugar production. The results suggest that the SH-proteinase is intimately involved in the mobilization of storage proteins.

  19. Effects of leupeptin on proteinase and germination of castor beans.

    PubMed

    Alpi, A; Beevers, H

    1981-10-01

    Leupeptin, a tripeptide inhibitor of some proteinases, was shown previously to maintain the stability of several enzymes (isocitrate lyase, fumarase, and catalase) in crude extracts of castor bean endosperm. This reagent is now shown to inhibit the breakdown of water-soluble and crystalloidstorage proteins of the protein bodies isolated from castor beans by the SH-proteinase and it also inhibits the endopeptidase from mung beans. When suitably introduced into the endosperm of dry castor beans it strongly inhibits germination and seedling development. Application of leupeptin to endosperm halves removed from the seed prevents the normal development of enzymes concerned with gluconeogenesis from fat and drastically curtails sugar production. The results suggest that the SH-proteinase is intimately involved in the mobilization of storage proteins.

  20. Multisystem organ failure after large volume injection of castor oil.

    PubMed

    Smith, Silas W; Graber, Nathan M; Johnson, Rudolph C; Barr, John R; Hoffman, Robert S; Nelson, Lewis S

    2009-01-01

    We report a case of multisystem organ failure after large volume subcutaneous injection of castor oil for cosmetic enhancement. An unlicensed practitioner injected 500 mL of castor oil bilaterally to the hips and buttocks of a 28-year-old male to female transsexual. Immediate local pain and erythema were followed by abdominal and chest pain, emesis, headache, hematuria, jaundice, and tinnitus. She presented to an emergency department 12 hours postinjection. Persistently hemolyzed blood samples complicated preliminary laboratory analysis. She rapidly deteriorated despite treatment and developed fever, tachycardia, hemolysis, thrombocytopenia, hepatitis, respiratory distress, and anuric renal failure. An infectious diseases evaluation was negative. After intensive supportive care, including mechanical ventilation and hemodialysis, she was discharged 11 days later, requiring dialysis for an additional 1.5 months. Castor oil absorption was inferred from recovery of the Ricinus communis biomarker, ricinine, in the patient's urine (41 ng/mL). Clinicians should anticipate multiple complications after unapproved methods of cosmetic enhancement.

  1. CASTOR: Cathode/Anode Satellite Thruster for Orbital Repositioning

    NASA Technical Reports Server (NTRS)

    Mruphy, Gloria A.

    2010-01-01

    The purpose of CASTOR (Cathode/Anode Satellite Thruster for Orbital Repositioning) satellite is to demonstrate in Low Earth Orbit (LEO) a nanosatellite that uses a Divergent Cusped Field Thruster (DCFT) to perform orbital maneuvers representative of an orbital transfer vehicle. Powered by semi-deployable solar arrays generating 165W of power, CASTOR will achieve nearly 1 km/s of velocity increment over one year. As a technology demonstration mission, success of CASTOR in LEO will pave the way for a low cost, high delta-V orbital transfer capability for small military and civilian payloads in support of Air Force and NASA missions. The educational objective is to engage graduate and undergraduate students in critical roles in the design, development, test, carrier integration and on-orbit operations of CASTOR as a supplement to their curricular activities. This program is laying the foundation for a long-term satellite construction program at MIT. The satellite is being designed as a part of AFRL's University Nanosatellite Program, which provides the funding and a framework in which student satellite teams compete for a launch to orbit. To this end, the satellite must fit within an envelope of 50cmx50cmx60cm, have a mass of less than 50kg, and meet stringent structural and other requirements. In this framework, the CASTOR team successfully completed PDR in August 2009 and CDR in April 2010 and will compete at FCR (Flight Competition Review) in January 2011. The complexity of the project requires implementation of many systems engineering techniques which allow for development of CASTOR from conception through FCR and encompass the full design, fabrication, and testing process.

  2. Large nonlocal nonlinear optical response of castor oil

    NASA Astrophysics Data System (ADS)

    Souza, Rogério F.; Alencar, Márcio A. R. C.; Meneghetti, Mario R.; Hickmann, Jandir M.

    2009-09-01

    The nonlocal nonlinearity of castor oil was investigated using the Z-scan technique in the CW regime at 514 nm and in femtosecond regime at 810 nm. Large negative nonlinear refractive indexes of thermal origin, thermo-optical coefficients and degree of nonlocality were obtained for both laser excitation wavelengths. The results indicate that the electronic part of the nonlinear refractive index and nonlinear absorption were negligible. Our results suggest that castor oil is promising candidate as a nonlinear medium for several nonlocal optical applications, such as in spatial soliton propagation, as well as a dispersant agent in the measurement of absorptive properties of nanoparticles.

  3. Synthesis of carbon nanotubes using natural carbon precursor: Castor oil

    NASA Astrophysics Data System (ADS)

    Raziah, A. Z.; Junizah, A. R.; Saifuddin, N.

    2012-09-01

    Castor oil has long been an article of commerce due to its versatility as it is widely used as a starting material for many industrial chemical products because of its unique structure. In this study, carbon nanotubes has been synthesized by thermal decomposition of castor oil in nitrogen atmosphere at 300-400δC using custom-made microwave processing unit. The precursor material was catalyzed by iron clusters originating from the addition of ferrocene. The morphology and characterization of the CNTs were studied and discussed by transmission electron microscopy (TEM).

  4. Laser-fabricated castor oil-capped silver nanoparticles.

    PubMed

    Zamiri, Reza; Zakaria, Azmi; Abbastabar, Hossein; Darroudi, Majid; Husin, Mohd Shahril; Mahdi, Mohd Adzir

    2011-01-01

    Silver nanoparticles were fabricated by ablation of a pure silver plate immersed in castor oil. A Nd:YAG-pulsed Q-switch laser with 1064-nm wavelength and 10-Hz frequency was used to ablate the plate for 10 minutes. The sample was characterized by ultraviolet-visible, atomic absorption, Fourier transform-infrared spectroscopies, and transmission electron microscopy. The results of the fabricated sample showed that the nanoparticles in castor oil were about 5-nm in diameter, well dispersed, and showed stability for a long period of time.

  5. IN VITRO REGENERATION CASTOR (RICINUS COMMUNIS L.) USING COTYLEDON EXPLANTS

    USDA-ARS?s Scientific Manuscript database

    A novel plant regeneration protocol was established for castor (Ricinus communis L.), an important oilseed crop. Mature seed-derived cotyledon explants produced adventitious shoots when placed on Murashige and Skoog (MS) medium containing thidiazuron (TDZ). The rate of shoot regeneration was maximal...

  6. Evaluation of castor oil samples for potential toxin contamination

    USDA-ARS?s Scientific Manuscript database

    Castor oil and its derivatives are widely used as a chemical feedstock for production of lubricants and greases, and for engineering plastics, plasticizers and surfactants. They also have wide application in consumer goods such as lipstick, deodorants and medicinal products. Due to concerns about th...

  7. Detoxification of castor meal through reactive seed crushing

    USDA-ARS?s Scientific Manuscript database

    Non-edible oil crops, such as castor or jatropha, contain several toxic components. Post-harvest treatments should be used to reduce the risks associated with the possible dispersion of toxic compounds in the environment. A new processing technology named Reactive Seed Crushing was developed, which ...

  8. Purification of leucoplast pyruvate kinase from developing castor bean endosperm.

    PubMed

    Plaxton, W C; Dennis, D T; Knowles, V L

    1990-12-01

    Leucoplast pyruvate kinase from endosperm of developing castor oil seeds (Ricinus communis L.; cv Baker) has been purified 1370-fold to a specific activity of 41.1 micromoles pyruvate produced per minute per milligram protein. Nondenaturing polyacrylamide gel electrophoresis of the purified enzyme resulted in a single protein staining band that co-migrated with pyruvate kinase activity. However, following sodium dodecyl sulfate polyacrylamide electrophoresis, two major protein staining bands of 57.5 and 44 kilodaltons, which occurred in an approximate 2:1 ratio, respectively, were observed. The native molecular mass was approximately 305 kilodaltons. Rabbit antiserum raised against the final enzyme preparation effectively immunoprecipitated leucoplast pyruvate kinase. The 57.5- and 44-kilodalton polypeptides are immunologically related as both proteins cross-reacted strongly on Western blots probed with the rabbit anti-(developing castor seed endosperm leucoplast pyruvate kinase) immunoglobulin that had been affinity-purified against the 57.5-kilodalton polypeptide. In contrast, pyruvate kinases from the following sources showed no immunological cross-reactivity with the same immunoglobulin: the cytosolic enzyme from developing or germinating castor bean endosperm; chloroplastic pyruvate kinase from expanding leaves of the castor oil plant; chloroplastic or cytosolic pyruvate kinase from the green alga, Selenastrum minutum; and mammalian or bacterial pyruvate kinases.

  9. A review on the challenges for increassed production of castor

    USDA-ARS?s Scientific Manuscript database

    The oil produced by castor (Ricinus communis) is an important raw material for chemical industry for its unique properties such as the only commercial source of a hydroxilated fatty acid that composes around 90% of the oil. This crop has a remarkable potential for increasing importance as a raw mate...

  10. Composition of acylglycerols in castor oil and their biosynthetic pathway

    USDA-ARS?s Scientific Manuscript database

    Castor oil has many industrial uses, such as the manufacture of aviation lubricant, plastics, paints, coatings, and cosmetics, because of its high content of ricinoleate (hydroxy fatty acid). We have identified many molecular species of acylglycerols using electrospray ionization mass spectrometry o...

  11. Trial of an experimental castor oil solution for cleaning dentures.

    PubMed

    Andrade, Ingrid Machado de; Andrade, Kelly Machado de; Pisani, Marina Xavier; Silva-Lovato, Cláudia Helena; de Souza, Raphael Freitas; Paranhos, Helena de Freitas Oliveira

    2014-01-01

    Denture hygiene is essential because denture biofilm is involved in oral infections and systemic diseases. Although there are chemical agents available on the market, none of them have ideal properties and research on such products is still necessary. The aim of this study was to evaluate the efficacy of a castor bean (Ricinus communis)-based solution for removing denture biofilm, compared to two traditional products (sodium hypochlorite and alkaline peroxide). Fifty maxillary complete denture wearers were instructed to brush their dentures after meals and to immerse their dentures once a day in the following solutions: Saline (20 min; control), Polident alkaline peroxide (3 min), NaOCl (20 min) and 2% castor oil solution (20 min). Participants used each solution for a period of 7 consecutive days, according to a random sequence. After each period, the internal surfaces of maxillary complete dentures were stained with a disclosing solution (1% neutral red), photographed and the disclosed biofilm was quantified with the aid of specific software. The influence of treatments on results was verified by the Friedman test (α=0.05). Tested solutions presented significant difference (Fr=51.67; p<0.001). Saline and NaOCl were significantly different (median: 2.0% and 0.0%) whereas Polident and castor oil presented intermediate results (median: 1.0% and 1.5%, respectively). It can be concluded that the castor oil solution tested in this study was comparable to alkaline peroxide in terms of efficiency in denture biofilm removal.

  12. Develop a New Lesquerella fendleri Crop for Castor Oil Production

    USDA-ARS?s Scientific Manuscript database

    Seed oil of Lesquerella fendleri contains a valuable hydroxy fatty acid (HFA), lesquerolic acid (20:1OH). The conventional source of HFA is ricinoleic acid (18:1OH) from castor seeds. Ricinoleic acid and its derivatives are used as raw materials for numerous industrial products, such as lubricants, ...

  13. GNS Castor V/21 Headspace Gas Sampling 2014

    SciTech Connect

    Winston, Philip Lon

    2016-01-01

    Prior to performing an internal visual inspection, samples of the headspace gas of the GNS Castor V/21 cask were taken on June 12, 2014. These samples were taken in support of the CREIPI/Japanese nuclear industry effort to validate fuel integrity without visual inspection by measuring the 85Kr content of the cask headspace

  14. Identification of minor acylglycerols less polar than triricinolein in castor oil by mass spectrometry and the proposed biosynthetic pathway of castor oil

    USDA-ARS?s Scientific Manuscript database

    Ricinoleate (OH18:1), a monohydroxy fatty acid, has many industrial uses such as the manufacture of biodegradable plastics, nylon, plasticizers, lubricants, cosmetics and paints. Castor oil is the only commericial source of ricinoleate which occurs as triacylglycerols. Triacylglycerols in castor oil...

  15. Role of tachykinins in castor oil diarrhoea in rats.

    PubMed

    Croci, T; Landi, M; Emonds-Alt, X; Le Fur, G; Maffrand, J P; Manara, L

    1997-06-01

    1. We set out to ascertain the role of tachykinins, neurokinin A and substance P, in castor oil-induced diarrhoea in rats as disclosed by the inhibitory effect of the non-peptide NK1- and NK2-receptor antagonists. SR 140333 and SR 48968, respectively. 2. SR 48968 (0.02 to 20 micrograms kg-1, s.c. or p.o.), and the opioid receptor agonist loperamide (1-10 mg kg-1, p.o.), dose-dependently prevented castor oil effects: % inhibition vs castor oil, diarrhoea 0 to 100, increase in faecal mass 7 to 90 and water content 16 to 90. SR 140333 (0.02 to 20 micrograms kg-1, s.c.) and the platelet activating factor antagonist SR 27417 (5 to 500 micrograms kg-1, p.o.) did not prevent the increase in faecal water content, but reduced faecal mass (35 to 66%) and diarrhoea (0 to 57%). 3. The R-enantiomers of tachykinin NK1 and NK2 receptor antagonists, SR 140603 and SR 48605 (both at 2 or 20 micrograms kg-1, s.c.) had no effect other than reducing faecal mass at the highest dose tested. 4. SR 48968 (20 micrograms kg-1, p.o.) but not loperamide (10 mg kg-1, p.o.) given 24 h before castor oil, still slightly but significantly reduced by 30% the increase of faecal mass output; both treatments significantly reduced (30 to 70%) the effect of castor oil on faecal water content, although the incidence of diarrhoea was only slightly less than in controls. 5. In castor oil-treated rats, naloxone (2 mg kg-1, s.c.) completely blocked the antidiarrhoeal action of loperamide (10 mg kg-1, p.o.) but not of SR 48968 (20 micrograms kg-1, p.o.): a similar result was obtained on faecal mass and water content. 6. Castor oil strongly increased the occurrence of manometrically recorded propulsive giant contractions (500 to 1000% over control values) of transverse and distal colon, this effect being significantly prevented (80 to 100%) by SR 48968 and loperamide and partially by SR 140333 (35% distal colon, 70% transverse colon). 7. In castor oil free rats, loperamide but not SR 48968 or SR 140333

  16. Draft genome sequence of the ricin-producing oilseed castor bean

    PubMed Central

    Chan, Agnes P.; Crabtree, Jonathan; Zhao, Qi; Lorenzi, Hernan; Orvis, Joshua; Puiu, Daniela; Melake-Berhan, Admasu; Jones, Kristine M.; Redman, Julia; Chen, Grace; Cahoon, Edgar B.; Gedil, Melaku; Stanke, Mario; Haas, Brian J.; Wortman, Jennifer R.; Fraser-Liggett, Claire M.; Ravel, Jacques; Rabinowicz, Pablo D.

    2010-01-01

    Castor bean (Ricinus communis) is an oil crop that belongs to the spurge (Euphorbiaceae) family. Its seeds are the source of castor oil, used for the production of high-quality lubricants due to its high proportion of the unusual fatty acid ricinoleic acid. Castor bean seeds also produce ricin, a highly toxic ribosome inactivating protein, making castor bean relevant for biosafety. We report here the 4.6X draft genome sequence of castor bean, representing the first reported Euphorbiaceae genome sequence. Our analysis shows that most key castor oil metabolism genes are single-copy while the ricin gene family is larger than previously thought. Comparative genomics analysis suggests the presence of an ancient hexaploidization event that is conserved across the dicotyledonous lineage. PMID:20729833

  17. An examination of the effect of castor oil packs on constipation in the elderly.

    PubMed

    Arslan, Gülşah Gürol; Eşer, Ismet

    2011-02-01

    This research, conducted at two rest homes in Manisa, Turkey, was undertaken to examine the effect of castor oil pack (COP) administrations on constipation in the elderly. Study participants were monitored for 7 days before, 3 days during, and 4 days after COP administration utilizing the Recall Bias and Visual Scale Analog (RB-VSAQ) as well as the Standard Diary developed by Pamuk et al. Wilcoxon Signed Ranks, Repeated Measures, Bonferroni, and Mann-Whitney U tests were used for data analysis. Eighty percent of study subjects had been constipated for 10 years or longer. COP administration did not have an effect on the number of bowel movements or amount of feces, but decreased the feces consistency score, straining during defecation and feeling of complete evacuation after a bowel movement, thus decreasing symptoms of constipation. We conclude that COP may be used for controlling symptoms of constipation.

  18. Development of a novel set of microsatellite markers for castor bean, Ricinus communis (Euphorbiaceae).

    PubMed

    Bajay, Miklos Maximiliano; Zucchi, Maria Imaculada; Kiihl, Tammy Aparecida Manabe; Batista, Carlos Eduardo Araújo; Monteiro, Mariza; Pinheiro, José Baldin

    2011-04-01

    Microsatellite primers were developed for castor bean (Ricinus communis L.) to investigate genetic diversity and population structure, and to provide support to germplasm management. Eleven microsatellite loci were isolated using an enrichment cloning protocol and used to characterize castor bean germplasm from the collection at the Instituto Agronômico de Campinas (IAC). In a survey of 76 castor bean accessions, the investigated loci displayed polymorphism ranging from two to five alleles. The information derived from microsatellite markers led to significant gains in conserved allelic richness and provides support to the implementation of several molecular breeding strategies for castor bean.

  19. Castor bean toxicity re-examined: a new perspective.

    PubMed

    Rauber, A; Heard, J

    1985-12-01

    Commonly used references present a very gloomy prognosis for castor seed ingestion. This appears to be based chiefly on extrapolations from laboratory animal studies of toxicity of the component, ricin. This is not in concordance with actual experience when actual case report experience is sought out and examined in detail. The literature has been confused by recursive references among secondary sources and mortality statistics have been distorted by reliance on data generated nearly a century ago which do not accurately reflect the capabilities of modern medical management. This is elucidated by actually tracking down and studying difficult-to-find original case reports from the early decades of the 20th century. It is time to reappraise our estimates of the hazard of ingesting castor seeds in the light of present day medical capabilities.

  20. "Castor Oil" - The Culprit of Acute Hair Felting.

    PubMed

    Maduri, V Ramya; Vedachalam, Ahalya; Kiruthika, S

    2017-01-01

    Acute hair felting is a rare disorder of scalp hair. In this condition, the hair becomes twisted, entangled as a hard stony mass resembling a bird's nest. Sudden hair matting has been reported earlier in the literature after vigorous use of chemical and herbal shampoos. Plica polonica is a patchy area of hair matting occurring in due course of time in neglected hair or underlying psychiatric illness. This case is interesting as the whole scalp hair matted immediately after using coconut oil and castor oil following washing. Growing long hair and taking oil bath are cultural and religious customs in South India. The high viscosity of castor oil and long hair had contributed to sudden felting of hair. This disorder of hair is irreversible and the hair should be cut off. Acute nature of this disorder will result in a serious psychological impact on the patient and the family.

  1. Experimental habitat scenting inhibits colonization by beaver,Castor canadensis.

    PubMed

    Gregory Welsh, R; Muller-Schwarze, D

    1989-03-01

    Unoccupied beaver (Castor canadensis) sites in New York State were for two years experimentally scented with a mixture of beaver castoreum and anal gland secretion. These sites were colonized less often than unscented control sites. The beaver is the first mammal to have been shown experimentally to use intraspecific odor cues when settling in vacant habitat. Territorial pheromones may be useful as repellents for beaver or other rodents.

  2. Forward Physics with the CMS-CASTOR calorimeter

    NASA Astrophysics Data System (ADS)

    van Haevermaet, Hans

    2010-02-01

    The CASTOR calorimeter, located at 14.4 from the CMS interaction point, enhances the hermeticity of the CMS detector by extending the rapidity coverage to η< 6.6. After having described the CASTOR calorimeter, examples of the capabilities and physics program of this detector will be presented. The study of forward jets production will give access to the low x > 10-6 region where the gluon distribution increase is expected to saturate through the effects of nonlinear recombination terms in the DGLAP evolution equation. The study of events with a centrally produced dijet system and a forward jet will give access to parton dynamics beyond DGLAP and to the investigation of the BFKL or BFKL-like QCD evolution. The study of forward - central multiplicity correlations will enable to obtain a better understanding of the theoretically poorly constrained soft background that are the underlying events, defined as everything except the hard scattered components. The use of CASTOR as forward activity veto detectors is also of prime importance in the study of diffractive events, such as single diffractive W production, in order to obtain a more powerful rejection of the non-diffractive and inelastic diffractive backgrounds. )

  3. Castor bean (Ricinus communis L.) as a potential environmental bioindicator.

    PubMed

    Mendes, M G; Santos Junior, C D; Dias, A C C; Bonetti, A M

    2015-10-21

    Biomonitoring of air quality using living organisms is a very interesting approach to environmental impact assessment. Organisms with a vast distribution, such as plants, are widely used for these purposes. The castor bean (Ricinus communis L.) is an oleaginous plant that can potentially be used as a bioindicator plant owing to its rapid growth and large leaves, which have a wide surface area of contact with the air and the pollutants therein. This study investigated the the bioindicator potential of the castor bean by performing several tests. We observed statistically significant differences in the concentrations of chlorophyll a and b in the leaves of plants in polluted areas compared to that in the control group plants, which were located in a pollution-free area. Leaves of plants in the former group had higher peroxidase activity and showed a greater buffering ability than those of plants in the control group. The pKa values obtained via buffering capacity tests, revealed the presence of aminoazobenzene (an industrial dye) in leaves of R. communis. Genotoxicity was evaluated through the comet assay technique and revealed that other than some differences in DNA fragmentation, there is no statistically significant difference in this parameter between places analyzed. Our data indicate that R. communis can be a highly useful biological indicator. Further, we hypothesized that the castor bean can be a potential candidate for phytoremediation owing its physiological buffering capacity when exposed to substantial pollution.

  4. Low-modulus PMMA bone cement modified with castor oil.

    PubMed

    López, Alejandro; Hoess, Andreas; Thersleff, Thomas; Ott, Marjam; Engqvist, Håkan; Persson, Cecilia

    2011-01-01

    Some of the current clinical and biomechanical data suggest that vertebroplasty causes the development of adjacent vertebral fractures shortly after augmentation. These findings have been attributed to high injection volumes as well as high Young's moduli of PMMA bone cements compared to that of the osteoporotic cancellous bone. The aim of this study was to evaluate the use of castor oil as a plasticizer for PMMA bone cements. The Young's modulus, yield strength, maximum polymerization temperature, doughing time, setting time and the complex viscosity curves during curing, were determined. The cytotoxicity of the materials extracts was assessed on cells of an osteoblast-like cell line. The addition of up to 12 wt% castor oil decreased yield strength from 88 to 15 MPa, Young's modulus from 1500 to 446 MPa and maximum polymerization temperature from 41.3 to 25.6°C, without affecting the setting time. However, castor oil seemed to interfere with the polymerization reaction, giving a negative effect on cell viability in a worst-case scenario.

  5. Measurement of quadratic electrogyration effect in castor oil

    NASA Astrophysics Data System (ADS)

    Izdebski, Marek; Ledzion, Rafał; Górski, Piotr

    2015-07-01

    This work presents a detailed analysis of electrogyration measurement in liquids with the usage of an optical polarimetric technique. Theoretical analysis of the optical response to an applied electric field is illustrated by experimental data for castor oil which exhibits natural optical activity, quadratic electro-optic effect and quadratic electrogyration effect. Moreover, the experimental data show that interaction of the oil with a pair of flat electrodes induces a significant dichroism and natural linear birefringence. The combination of these effects occurring at the same time complicates the procedure of measurements. It has been found that a single measurement is insufficient to separate the contribution of the electrogyration effect, but it is possible on the basis of several measurements performed with various orientations of the polarizer and the analyser. The obtained average values of the quadratic electrogyration coefficient β13 in castor oil at room temperature are from - 0.92 ×10-22 to - 1.44 ×10-22m2V-2 depending on the origin of the oil. Although this study is focused on measurements in castor oil, the presented analysis is much more general.

  6. Fatty acids in extremity tissues of Finnish beavers (Castor canadensis and Castor fiber) and muskrats (Ondatra zibethicus).

    PubMed

    Käkelä, R; Hyvärinen, H

    1996-01-01

    Site-specific differences in fatty acid composition of adipose tissues from Finnish beavers (Castor canadensis and Castor fiber) and muskrats (Ondatra zibethicus) were studied by gas liquid chromatography. In the total lipids of heterothermic extremities the proportions of potentially endogenous (delta 9-desaturated) monoenoic fatty acids were 1.5-fold and those of the saturated fatty acids 0.5-fold the proportions of subcutaneous or inner depots. Thus the role of delta 9-desaturation in the biochemical adaptation of the extremity tissues is discussed. Dietary polyunsaturated fatty acids, with the exception of 18:3 omega 3 in the beavers, did not accumulate selectively in the extremity tissues. The liver of the muskrat (which also eats fish and mussels) contained high proportions of 22:6 omega 3, which was a minor component in the beaver (exclusively a herbivore).

  7. Demodex castoris sp. nov. (Acari: Demodecidae) parasitizing Castor fiber (Rodentia), and other parasitic arthropods associated with Castor spp.

    PubMed

    Izdebska, Joanna N; Fryderyk, Sławomira; Rolbiecki, Leszek

    2016-02-11

    A new species of demodecid mite, Demodex castoris sp. nov. (Acari: Prostigmata: Demodecidae), is described based on adult stages from the skin of the nasal region of the Eurasian beaver Castor fiber Linnaeus, 1758, collected in Poland. This is the first detection of a representative demodecid mite in rodents of the suborder Castorimorpha and also represents the first detection of a skin mite in Eurasian beavers. The new species is a small skin mite (average 173 µm in length) characterized by sexual dimorphism related to body proportions. D. castoris sp. nov. was observed in 4 out of 6 beavers examined (66.6%), with a mean intensity of 10.8 and an intensity range of 2-23 ind. host(-1). This paper also contains a checklist of parasitic arthropods known from Castor spp.

  8. Dirt-binding particles consisting of hydrogenated castor oil beads constitute a nonirritating alternative for abrasive cleaning of recalcitrant oily skin contamination in a three-step programme of occupational skin protection.

    PubMed

    Mahler, V; Erfurt-Berge, C; Schiemann, S; Michael, S; Egloffstein, A; Kuss, O

    2010-04-01

    In occupational fields with exposure to grease, oil, metal particles, coal, black lead or soot, cleansing formulations containing abrasive bodies (e.g. refined walnut shell, corn, wood, plastic or pumice) are used. These may constitute an irritant per se. As an alternative, hydrogenated castor oil (also known as castor wax) beads have been developed as dirt-binding particles. A polar surface contributes to their mechanical cleaning effects in removal of oily grime. Standardized examination of the in vivo effects upon the skin barrier of castor wax beads in comparison with abrasive bodies and pure detergent. Three cleansing preparations - (i) detergent, (ii) detergent containing castor wax beads, (iii) detergent containing walnut shell powder - were each repetitively applied in vivo (four times daily for 3 weeks), mimicking workplace conditions, in 30 healthy volunteers (15 with and 15 without an atopic skin diathesis) and compared vs. (iv) no treatment. The treatment effects upon the skin barrier were monitored by repeated measurements of functional parameters [transepidermal water loss (TEWL), redness] and surface topography. After a 3-week treatment, a significant global treatment effect (P < 0.0001) was found in the atopic group concerning TEWL as indicator for barrier function. A significantly higher TEWL and increasing erythema in the area treated with detergent containing walnut shell powder reflected its irritant effect compared with castor wax beads dispensed in the identical detergent. Cleaning properties of the two formulas were comparably superior to detergent alone. Castor wax beads constitute a novel nonirritating alternative for abrasive cleaning of recalcitrant oily skin contamination appropriate for individuals with an atopic skin diathesis in a three-step programme of occupational skin protection. As the skin barrier may additionally be influenced by the composition of dirt and use of skin protection and skin care measures under real workplace

  9. Mass spectrometry of the lithium adducts of diacylglycerols containing hydroxy FA in castor oil and two normal FA

    USDA-ARS?s Scientific Manuscript database

    Castor oil can be used in industry. The molecular species of triacylglycerols containing hydroxy fatty acids (FA) in castor oil have been identified. We report here the identification of twelve diacylglycerols (DAG) containing hydroxy FA in castor oil using positive ion electrospray ionization mass ...

  10. Identification of diacylglycerol and triacylglycerol containing 11,12,13-trihydroxy-9-14-octadecadienoic acids in castor oil

    USDA-ARS?s Scientific Manuscript database

    Castor oil has many industrial uses. Molecular species of acylglycerols containing monohydroxy, dihydroxy and trihydroxy fatty acids in castor oil have been reported. We report here the identification of acylglycerols containing triOH18:2 fatty acid in castor oil. The structure of this novel fa...

  11. Identification of minor acylglycerols less polar than triricinolein in castor oil by mass spectrometry

    USDA-ARS?s Scientific Manuscript database

    Triacylglycerols in castor oil less polar than triricinolein were identified by electrospray ionization – mass spectrometry using the lithium adducts of the acylglycerols from the HPLC fractions of castor oil. Thirty four new molecular species of acylglycerols containing hydroxy fatty acids in casto...

  12. Identification of acylglycerols containing dihydroxy fatty acids in castor oil by mass spectrometry

    USDA-ARS?s Scientific Manuscript database

    Ricinoleate, a monohydroxy fatty acid, in castor oil has many industrial uses. Dihydroxy fatty acids can also be used in industry. The C18 HPLC fractions of castor oil were used for mass spectrometry of lithium addicts to identify the acylglycerols containing dihydroxy fatty acids. Four diacylglycer...

  13. Expression profile of the genes involved in reserve synthesis in castor (Ricinus communis L.)

    USDA-ARS?s Scientific Manuscript database

    Oil derived from the seed of castor (Ricinus communis L.) contains 90% ricinoleate (12-hydroxy-oleate) and has numerous industrial uses. Despite its economic importance, the production of castor oil is hampered by the presence of detrimental seed storage proteins, the toxin ricin and hyper-allergeni...

  14. Seed Development in Castor (Ricinus communis L.): Morphology, Reserve Synthesis and Gene Expression

    USDA-ARS?s Scientific Manuscript database

    Castor (Ricinus communis L.) is a non-eatable oilseed crop producing seed oil comprising 90% ricinoleate (12-hydroxy-oleate) which has numerous industrial uses. However, the production of castor oil is hampered by the presence of noxious seed storage proteins, the toxin ricin and hyper-allergenic 2S...

  15. Oil content among the diverse castor genetic resources in the U.S. collection

    USDA-ARS?s Scientific Manuscript database

    Castor (Ricinus communis L.) contains oil used for medicine, as an ingredient in shampoo, soap, hand lotion, high-speed lubricants, and as a coating material. Due to its high oil content, oil derived from castor seeds is currently proposed to be used as a feedstock for biodiesel production. The USDA...

  16. Castor oil induces laxation and uterus contraction via ricinoleic acid activating prostaglandin EP3 receptors

    PubMed Central

    Tunaru, Sorin; Althoff, Till F.; Nüsing, Rolf M.; Diener, Martin; Offermanns, Stefan

    2012-01-01

    Castor oil is one of the oldest drugs. When given orally, it has a laxative effect and induces labor in pregnant females. The effects of castor oil are mediated by ricinoleic acid, a hydroxylated fatty acid released from castor oil by intestinal lipases. Despite the wide-spread use of castor oil in conventional and folk medicine, the molecular mechanism by which ricinoleic acid acts remains unknown. Here we show that the EP3 prostanoid receptor is specifically activated by ricinoleic acid and that it mediates the pharmacological effects of castor oil. In mice lacking EP3 receptors, the laxative effect and the uterus contraction induced via ricinoleic acid are absent. Although a conditional deletion of the EP3 receptor gene in intestinal epithelial cells did not affect castor oil-induced diarrhea, mice lacking EP3 receptors only in smooth-muscle cells were unresponsive to this drug. Thus, the castor oil metabolite ricinoleic acid activates intestinal and uterine smooth-muscle cells via EP3 prostanoid receptors. These findings identify the cellular and molecular mechanism underlying the pharmacological effects of castor oil and indicate a role of the EP3 receptor as a target to induce laxative effects. PMID:22615395

  17. Castor oil induces laxation and uterus contraction via ricinoleic acid activating prostaglandin EP3 receptors.

    PubMed

    Tunaru, Sorin; Althoff, Till F; Nüsing, Rolf M; Diener, Martin; Offermanns, Stefan

    2012-06-05

    Castor oil is one of the oldest drugs. When given orally, it has a laxative effect and induces labor in pregnant females. The effects of castor oil are mediated by ricinoleic acid, a hydroxylated fatty acid released from castor oil by intestinal lipases. Despite the wide-spread use of castor oil in conventional and folk medicine, the molecular mechanism by which ricinoleic acid acts remains unknown. Here we show that the EP(3) prostanoid receptor is specifically activated by ricinoleic acid and that it mediates the pharmacological effects of castor oil. In mice lacking EP(3) receptors, the laxative effect and the uterus contraction induced via ricinoleic acid are absent. Although a conditional deletion of the EP(3) receptor gene in intestinal epithelial cells did not affect castor oil-induced diarrhea, mice lacking EP(3) receptors only in smooth-muscle cells were unresponsive to this drug. Thus, the castor oil metabolite ricinoleic acid activates intestinal and uterine smooth-muscle cells via EP(3) prostanoid receptors. These findings identify the cellular and molecular mechanism underlying the pharmacological effects of castor oil and indicate a role of the EP(3) receptor as a target to induce laxative effects.

  18. Methyl esters from vegetable oils with hydroxy fatty acids: Comparison of lesquerella and castor methyl esters

    USDA-ARS?s Scientific Manuscript database

    The search for alternative feedstocks for biodiesel as partial replacement for petrodiesel has recently extended to castor oil. In this work, the castor oil methyl esters were prepared and their properties determined in comparison to the methyl esters of lesquerella oil, which in turn is seen as alt...

  19. Genetic diversity and population structure of castor germplasm within the U.S. collection

    USDA-ARS?s Scientific Manuscript database

    Castor is an important oilseed crop and its oil is inedible but has multiple industrial and pharmaceutical applications. Castor germplasm collection is a useful genetic resource for improving seed quality and developing new cultivars. Seed oil content and fatty acid composition of the entire U.S. ca...

  20. Synthesis and membrane performance characterization of self-emulsified waterborne nitrocellulose dispersion modified with castor oil

    NASA Astrophysics Data System (ADS)

    Su, Xiuxia; Zhao, Qingxiao; Zhang, Dan; Dong, Wei

    2015-11-01

    Waterborne nitrocellulose dispersion modified with castor oil (CWNC) was designed and successfully synthesized by self emulsification and reaction among isophorone diisocyanate (IPDI) trimer, dimethylol propionic acid (DMPA), nitrocellulose (NC) and castor oil (C.O.). The CWNC was characterized by transmission electron microscope (TEM), Fourier transform infrared (FTIR) spectroscopy and thermal gravimetric analysis (TGA), etc. The particle size of CWNC increased with the increase of mass fraction of castor oil to total reactants, ω (C.O.). The morphology of particles is an approximate core-shell structure indicated by TEM. FTIR confirmed that the reactions (i.e. IPDI trimer and castor oil, IPDI trimer and NC) occurred, the NCO groups of IPDI trimer were consumed totally and the backbone of NC was retained. The water contact angle measurements confirmed that introduced castor oil increased hydrophobicity of the film, thereby increasing the contact angle. TGA revealed that the CWNC film had better thermal resistance.

  1. Physicochemical and sensory characteristics of milk from goats supplemented with castor or licuri oil.

    PubMed

    Pereira, R A G; Oliveira, C J B; Medeiros, A N; Costa, R G; Bomfim, M A D; Queiroga, R C R E

    2010-02-01

    The purpose of this study was to evaluate the influence of castor and licuri palm oils supplemented to milking goats on the physical, chemical, and sensory characteristics of milk. A double Latin square experimental design (5x5) using 10 confined crossbred Moxotó-Alpine goats was performed according to the following treatments: nonsupplemented (control), 3% castor oil, 5% castor oil, 3% licuri oil, and 5% licuri oil. Oils in each treatment were supplemented in the dry matter. Castor oil supplementation reduced the fat content and increased the lactose and density of milk. Considering the sensory analysis, a lower acceptability was observed for milk from goats supplemented with castor oil. On the other hand, licuri oil supplementation led to higher acceptability scores for flavor and odor of goat milk.

  2. Identification of diacylglycerol and triacylglycerol containing 11,12,13-trihydroxy-9,14-octadecadienoic acid in castor oil.

    PubMed

    Lin, Jiann-Tsyh; Chen, Grace Q

    2011-02-28

    Castor oil has many industrial uses. Molecular species of acylglycerols containing monohydroxy, dihydroxy and trihydroxy fatty acids in castor oil have been reported. We report here the identification of acylglycerols containing a triOH18:2 fatty acid in castor oil. The structure of this novel fatty acid was proposed as 11,12,13-trihydroxy-9,14-octadecadienoic acid by the mass spectrometry of the lithiated adducts of acylglycerols in the HPLC fractions of castor oil. The fragmentation pathways of the lithiated adduct of 11,12,13-trihydroxy-9,14-octadecadienoic acid were proposed. We also proposed the biosynthetic pathways of polyhydroxy fatty acids in castor.

  3. Castor bean (Ricinus communis) toxicosis in a sheep flock.

    PubMed

    Aslani, M R; Maleki, M; Mohri, M; Sharifi, K; Najjar-Nezhad, V; Afshari, E

    2007-03-01

    This paper describes clinical, laboratory and pathological findings of sheep, which is intoxicated with castor bean. The source of intoxication was a miscellaneous garden waste. Forty-five animals showed clinical toxicosis and 17 died. The clinical signs included weakness, salivation, profuse watery diarrhoea, dehydration, mydriasis, teeth grinding, hypothermia and recumbency. The most significant haematological and biochemical findings were a high haematocrit, high concentration of serum BUN, creatinine and phosphorus and high activity of serum CK and AST. Pathology revealed severe gastroenteritis, cardiac haemorrhage and necrosis, hepatic necrosis and acute tubular necrosis in kidneys. Treatment included symptomatic and supportive care with fluid therapy and cathartic administration.

  4. Investigation of Carbohydrate Metabolism and Transport in Castor Bean Seedlings by Cyclic JCross Polarization Imaging and Spectroscopy

    NASA Astrophysics Data System (ADS)

    Heidenreich, M.; Köckenberger, W.; Kimmich, R.; Chandrakumar, N.; Bowtell, R.

    1998-05-01

    NMR experiments using13C-labeled compounds offer the possibility of noninvasive monitoring of carbohydrate transport and metabolism in living plants, but are usually hampered by the low sensitivity of the13C nucleus. The problem of low sensitivity can be overcome by using the cyclicJcross polarization (CYCLCROP) technique, which allows the indirect detection of13C nuclei coupled to1H nuclei with the high NMR sensitivity of protons. We report here on methods for imaging and spectroscopy based on the CYCLCROP technique, and their use in the firstin vivoNMR study of carbohydrate transport and metabolism in castor bean seedlings (Ricinus communis L.). Comprehensive acquisition strategies for the various NMR methods are given, including the procedure for setting up the experiments. In addition, a full analysis of the effect of relaxation on the signals generated from smallJ-coupled spin systems by the CYCLCROP sequence is given, and the high sensitivity of the sequence is demonstrated. In thein vivostudy of six-day-old castor bean seedlings, we were able to measure the uptake of labeled hexoses, supplied in solution to the cotyledons, and their conversion to sucrose, as well as the transport of this sucrose in the vascular bundles. Images of the actual distribution of labeled sucrose in the hypocotyl of the seedling have also been obtained. The resulting data show some evidence for a preferential incorporation of labeled fructose in the process of sucrose synthesis, which decreases with the time of incubation.

  5. Spreader beam analysis for the CASTOR GSF cask

    SciTech Connect

    Clements, E.P.

    1997-04-07

    The purpose of this report is to document the results of the 150% rated capacity load test performed by DynCorp Hoisting and Rigging on the CASTOR GSF special cask lifting beams. The two lifting beams were originally rated and tested at 20,000kg (44,000lb) by the cask manufacturer in Germany. The testing performed by DynCorp rated and tested the lifting beams to 30,000 kg (66,000 lb) +0%, -5%, for Hanford Site use. The CASTOR GSF cask, used to transport isotopic Heat Sources (canisters), must be lifted with its own designed lifting beam system (Figures 1, 2, and 3). As designed, the beam material is RSt 37-2 (equivalent to American Society for Testing and Materials [ASTM] A-570), the eye plate is St 52-2 (equivalent to ASTM A-516), and the lifting pin is St 50 (equivalent to ASTM A-515). The beam has two opposing 58 mm (2.3 in.) diameter by 120 mm(4.7 in.) length, high grade steel pins that engage the cask for lifting. The pins have a manual locking mechanism to prevent disengagement from the casks. The static, gross weight (loaded) of the cask 18,640 kg (41,000 lb) on the pins prevents movement of the pins during lifting. This is due to the frictional force of the cask on the pins when lifting begins.

  6. Membrane Lipid Metabolism in Germinating Castor Bean Endosperm 1

    PubMed Central

    Donaldson, Robert P.

    1976-01-01

    Castor bean (Ricinus communis L. var. Hale) endosperms, excised after 2 days germination at 30 C, were incubated 5 min to 8 hr with 14C-acetate and 3H-glycerol. Homogenates were fractionated by sucrose gradient centrifugation. Organelles found to be active in lipid synthesis were the lipid bodies and the endoplasmic reticulum. The products of incorporation in the lipid bodies were 3H-diglycerides containing 14C-fatty acids of more than 20 carbons. In contrast, the endoplasmic reticulum produced 3H-phospholipids as well as 3H-diglycerides rich in 14C-linoleate. The phospholipids synthesized and their acyl contents were of the types known to be the major components of organelle membranes in this tissue. Phospholipids and diglycerides containing 14C and 3H were found in the glyoxysomes and mitochondria subsequent to their appearance in the endoplasmic reticulum. The results show that germinating castor bean endosperm synthesizes membrane lipids de novo from acetate rather than reutilizing stored lipid components directly. It is also apparent that the endoplasmic reticulum is responsible for several steps in membrane lipid production. PMID:16659516

  7. Production of ricinoleic acid from castor oil by immobilised lipases.

    PubMed

    Ozcan, Hakki Mevlut; Sagiroglu, Ayten

    2009-01-01

    Porcine pancreatic lipase (PPL), Candida rugosa lipase (CRL), and Castor bean lipase (CBL) were immobilized on celite by deposition from aqueous solution by the addition of hexane. Lipolytic performance of free and immobilized lipases were compared and optimizations of lipolytic enzymatic reactions conditions were performed by free and immobilized derivatives using olive oil as substrate. Afterwards, the influence on lipolysis of castor oil of free lipases and immobilized lipase derivatives have been studied in the case of production of ricinoleic acid. All of the lipases performances were compared and enzyme derivative was selected to be very effective on the production of ricinoleic acid by lipolysis reaction. Various reaction parameters affecting the production of ricinoleic acid were investigated with selected the enzyme derivative. The maximum ricinoleic acid yield was observed at pH 7-8, 50 degrees C, for 3 hours of reaction period with immobilized 1,3-specific PPL on celite. The kinetic constants K(m) and V(max) were calculated as 1.6 x 10(-4) mM and 22.2 mM from a Lineweaver-Burk plot with the same enzyme derivative. To investigate the operational stability of the lipase, the three step lipolysis process was repeated by transferring the immobilized lipase to a substrate mixture. As a result, the percentage of conversion after usage decreased markedly.

  8. Monitors.

    ERIC Educational Resources Information Center

    Powell, David

    1984-01-01

    Provides guidelines for selecting a monitor to suit specific applications, explains the process by which graphics images are produced on a CRT monitor, and describes four types of flat-panel displays being used in the newest lap-sized portable computers. A comparison chart provides prices and specifications for over 80 monitors. (MBR)

  9. Evaluation of the effect of castor oil on initiating labor in term pregnancy.

    PubMed

    Azhari, Sedighe; Pirdadeh, Sohaylla; Lotfalizadeh, Marzieh; Shakeri, Mohammad T

    2006-07-01

    To determine the effect of castor oil on initiating labor in term pregnant women. We conducted this randomized control clinical trial on 47 pregnant women from August 2003 to March 2004 in Mashhad University of Medical Sciences, Iran. There was a significant increase in labor initiation ratio in the castor oil group compared with the control group (54.2% compared with 4.3%, p<0.001). Also, the mean bishop score in the castor oil group increased from 2.50 +/- 1.29 to 6.79 +/- 3.20 (p<0.001). The probability of labor initiation increases during the first 24 hours after using castor oil, however, further studies are recommended to improve its efficacy and safety.

  10. Molecular species of acylglycerols incorporating radiolabeled fatty acids from castor (Ricinus communis L.) microsomal incubations.

    PubMed

    Lin, Jiann-Tsyh; Chen, Jennifer M; Liao, Lucy P; McKeon, Thomas A

    2002-08-28

    Sixty-one molecular species of triacylglycerols (TAG) and diacylglycerols produced from castor microsomal incubations incorporating six different (14)C-labeled fatty acids have been identified and quantified. The preference for incorporation into TAG was in the order ricinoleate > oleate > linoleate > linolenate > stearate > palmitate. Ricinoleate was the major fatty acid incorporated, whereas stearate, linolenate, and palmitate were incorporated at low levels. Twenty-one molecular species of acylglycerols (HPLC peaks) in castor oil have also been assigned. The levels of TAG in castor oil are RRR (triricinolein) > RR-TAG > R-TAG > no R-TAG. The levels of the molecular species within the groups of RR-TAG, RL-TAG, and LL-TAG individually are ricinoleate > linoleate > oleate > linolenate, stearate, and palmitate. The results of the labeled fatty acid incorporation are consistent with ricinoleate being preferentially driven into TAG and oleate being converted to ricinoleate in castor oil biosynthesis.

  11. Detection of acetone processing of castor bean mash for forensic investigation of ricin preparation methods.

    PubMed

    Kreuzer, Helen W; Wahl, Jon H; Metoyer, Candace N; Colburn, Heather A; Wahl, Karen L

    2010-07-01

    Samples containing the toxic castor bean protein ricin have been recently seized in connection with biocriminal activity. Analytical methods that enable investigators to determine how the samples were prepared and to match seized samples to potential source materials are needed. One commonly described crude ricin preparation method is acetone extraction of crushed castor beans. Here, we describe the use of solid-phase microextraction and headspace analysis to determine whether castor beans were processed by acetone extraction. We prepared acetone-extracted castor bean mash, along with controls of unextracted mash and mash extracted with nonacetone organic solvents. Samples of acetone-extracted mash and unextracted mash were stored in closed containers for up to 109 days at both room temperature and -20 degrees C, and in open containers at room temperature for up to 94 days. Acetone-extracted bean mash could consistently be statistically distinguished from controls, even after storage in open containers for 94 days.

  12. Red maple (Acer rubrum) inhibits feeding by beaver (Castor canadensis).

    PubMed

    Müller-Schwarze, D; Schulte, B A; Sun, L; Müller-Schwarze, A; Müller-Schwarze, C

    1994-08-01

    At many beaver (Castor canadensis) sites at Allegany State Park in New York State, red maple (Acer rubrum) is the only or one of the few tree species left standing at the ponds' edges. The relative palatability of red maple (RM) was studied in three ways. (1) At seven beaver sites, the available and utilized trees were recorded and an electivity index (E) computed. Of 15 tree species, RM ranked second or fourth lowest. (2) In experiment I, RM, sugar maple (A. saccharum, SM), and quaking aspen (Populus tremuloides) logs were presented cafeteria style at 10 colonies. RM was the least preferred. (3) Bark of RM was extracted with solvents. Aspen logs were painted (experiment II) or soaked (experiment III) with this RM extract and presented to beaver cafeteria-style, along with aspen and RM controls. This treatment rendered aspen logs less palatable, indicating that a chemical factor had been transferred.

  13. LDPE/PHB blends filled with castor oil cake

    NASA Astrophysics Data System (ADS)

    Burlein, Gustavo A.; Rocha, Marisa C. G.

    2015-05-01

    The response surface methodology (RSM) is a collection of mathematical techniques useful for developing, improving and optimizing process. In this study, RSM technique was applied to evaluate the effect of the components proportion on the mechanical properties of low density polyethylene (LDPE)/ poly (3-hydroxy-butyrate) (PHB) blends filled with castor oil cake (CC). The blends were prepared by melt mixing in a twin screw extruder. Low density polyethylene, poly (3-hydroxy-butyrate) and castor oil pressed cake were represented by the input variables designated as LDPE, PHB and CC, respectively. As it was desirable to consider the largest LDPE content in the ternary system, the components of the mixture were subjected to the following constraints: 0.7 ≤ LDPE ≤ 1.0, 0≤ PHB≤0.3 e 0 ≤ CC ≤0.3. The mechanical properties of the different mixtures were determined by conventional ASTM tests and were evaluated through analysis of variance performed by the Minitab software. Some polynomial equations were tested in order to describe the mechanical behavior of the samples. The quadratic model in pseudo components was selected for describing the tensile behavior because it was the most efficient from a statistical point of view (p-value ≤ 0.05; coefficient of determination (r2) close to 1 and variation inflation factor (VIF) values < 5). The results showed that the LDPE Young's modulus increases but the other tensile properties and impact resistance deteriorate with the addition of PHB or CC. The tensile strength values of binary mixtures of LDPE lie in the range from 8.9 to 10 MPa. As some commercial grades of LDPE have mechanical strength in this range, it may be inferred that the addition of a certain amount of PHB or CC to LDPE may be considered as a possibility for obtaining LDPE based materials with increased susceptibility to biodegradation. The cubic model in pseudo components was selected for describe the flexural strength of the samples because it was

  14. The chromosomes of the Canadian Beaver Castor canadensis.

    PubMed

    Genest, F B; Morisset, P; Patenaude, R P

    1979-03-01

    A chromosome analysis of 24 Canadian beavers, Castor canadensis Kuhl (12 males and 12 females), captured in Laurentides Park, Qébec, has been performed from preparations of blood lymphocyte and skin cultures. The chromosome number was found to be 2n = 40. Measurements were made to determine relative lengths and arm ratios of chromosomes, which are metacentric or submetacentric. Results are in agreement with those already published regarding the chromosome number, but differ in the identification of the X chromosome, and in the morphology of the Y and some autosomes. C- ad G-banding techniques allowed the precise identification of individual chromosome pairs. A detailed idiogram of G-bands is presented.

  15. Immobilization of Yarrowia lipolytica for aroma production from castor oil.

    PubMed

    Braga, Adelaide; Belo, Isabel

    2013-04-01

    The main aim of this study was to compare different materials for Y. lipolytica immobilization that could be used in the production of γ-decalactone (a peach-like aroma) in order to prevent the toxic effect both of the substrate and the aroma upon the cells. Therefore, cells adsorption onto pieces of methyl polymethacrylate and of DupUM(®) was studied and further used in the biotransformation of castor oil into γ-decalactone. The highest aroma concentration was obtained with immobilized cells in DupUM(®), where reconsumption of the aroma by the cells was prevented, contrarily to what happens with free cells. This is a very promising result for γ-decalactone production, with potential to be used at an industrial level since the use of immobilized cells system will facilitate the conversion of a batch process into a continuous mode keeping high cell density and allowing easier recovery of metabolic products.

  16. Castor oil polyurethane containing silica nanoparticles as filling material of bone defect in rats.

    PubMed

    Nacer, Renato Silva; Poppi, Rodrigo Ré; Carvalho, Paulo de Tarso Camilo de; Silva, Baldomero Antonio Kato da; Odashiro, Alexandre Nakao; Silva, Iandara Schettert; Delben, José Renato Jurkevicz; Delben, Angela Antonia Sanches Tardivo

    2012-01-01

    To evaluate the biologic behavior of the castor polymer containing silica nanoparticles as a bone substitute in diafisary defect. Twenty seven male Rattus norvegicus albinus Wistar lineage were submitted to bone defect filled with castor oil polymer. Three experimental groups had been formed with nine animals each: (1) castor oil polymer containing only calcium carbonate; (2) castor oil polymer with calcium carbonate and doped with 5% of silica nanoparticles; (3) castor polymer with calcium carbonate doped with 10% of silica nanoparticles; 3 animals of each group were submitted to euthanasia 15, 30 and 60 days after experimental procedure, and their femurs were removed to histological evaluation. there was bone growth in all the studied groups, with a greater tendency of growth in the group 1. After 30 days all the groups presented similar results. After 60 days a greater amount of fibroblasts, osteoblasts, osteocytes and osteoclasts in group 3 was observed, with integrated activity of 3 kinds of cells involved in the bone activation-reabsorption-formation. The castor polymer associated to the silica nanoparticles is biocompatible and allows osteoconduction. The presence of osteoprogenitors cells suggests silica osteoinduction capacity.

  17. CASTOR3D: linear stability studies for 2D and 3D tokamak equilibria

    NASA Astrophysics Data System (ADS)

    Strumberger, E.; Günter, S.

    2017-01-01

    The CASTOR3D code, which is currently under development, is able to perform linear stability studies for 2D and 3D, ideal and resistive tokamak equilibria in the presence of ideal and resistive wall structures and coils. For these computations ideal equilibria represented by concentric nested flux surfaces serve as input (e.g. computed with the NEMEC code). Solving an extended eigenvalue problem, the CASTOR3D code takes simultaneously plasma inertia and wall resistivity into account. The code is a hybrid of the CASTOR_3DW stability code and the STARWALL code. The former is an extended version of the CASTOR and CASTOR_FLOW code, respectively. The latter is a linear 3D code computing the growth rates of resistive wall modes in the presence of multiply-connected wall structures. The CASTOR_3DW code, and some parts of the STARWALL code have been reformulated in a general 3D flux coordinate representation that allows to choose between various types of flux coordinates. Furthermore, the implemented many-valued current potentials in the STARWALL part allow a correct treatment of the m  =  0, n  =  0 perturbation. In this paper, we outline the theoretical concept, and present some numerical results which illustrate the present status of the code and demonstrate its numerous application possibilities.

  18. Arsenic speciation in phloem and xylem exudates of castor bean.

    PubMed

    Ye, Wen-Ling; Wood, B Alan; Stroud, Jacqueline L; Andralojc, P John; Raab, Andrea; McGrath, Steve P; Feldmann, Jörg; Zhao, Fang-Jie

    2010-11-01

    How arsenic (As) is transported in phloem remains unknown. To help answer this question, we quantified the chemical species of As in phloem and xylem exudates of castor bean (Ricinus communis) exposed to arsenate [As(V)], arsenite [As(III)], monomethylarsonic acid [MMA(V)], or dimethylarsinic acid. In the As(V)- and As(III)-exposed plants, As(V) was the main species in xylem exudate (55%-83%) whereas As(III) predominated in phloem exudate (70%-94%). The ratio of As concentrations in phloem to xylem exudate varied from 0.7 to 3.9. Analyses of phloem exudate using high-resolution inductively coupled plasma-mass spectrometry and accurate mass electrospray mass spectrometry coupled to high-performance liquid chromatography identified high concentrations of reduced and oxidized glutathione and some oxidized phytochelatin, but no As(III)-thiol complexes. It is thought that As(III)-thiol complexes would not be stable in the alkaline conditions of phloem sap. Small concentrations of oxidized glutathione and oxidized phytochelatin were found in xylem exudate, where there was also no evidence of As(III)-thiol complexes. MMA(V) was partially reduced to MMA(III) in roots, but only MMA(V) was found in xylem and phloem exudate. Despite the smallest uptake among the four As species supplied to plants, dimethylarsinic acid was most efficiently transported in both xylem and phloem, and its phloem concentration was 3.2 times that in xylem. Our results show that free inorganic As, mainly As(III), was transported in the phloem of castor bean exposed to either As(V) or As(III), and that methylated As species were more mobile than inorganic As in the phloem.

  19. Development of nine new microsatellite loci for the American beaver, Castor canadensis (Rodentia: Castoridae), and cross-species amplification in the European beaver, Castor fiber

    USGS Publications Warehouse

    Pelz-Serrano, K.; Munguia-Vega, A.; Piaggio, A.J.; Neubaum, M.; Munclinger, P.; PArtl, A.; van Riper, Charles; Culver, M.

    2009-01-01

    We developed nine new nuclear dinucleotide microsatellite loci for Castor canadensis. All loci were polymorphic, except for one. The number of alleles ranged from two to four and from five to 12 in populations from Arizona and Wisconsin, respectively. Average heterozygosity ranged from 0.13 to 0.86 per locus. Since cross-species amplification in Castor fiber was successful only in four loci, we tested also nine recently published C. canadensis loci in the Eurasian species. Eight of the published loci amplified; however, three were monomorphic. The number of alleles was lower in C. fiber than in C. canadensis at all loci tested. ?? 2009 Blackwell Publishing Ltd.

  20. Screening of the entire USDA castor germplasm collection for oil content and fatty acid composition for optimum biodiesel production.

    PubMed

    Wang, Ming Li; Morris, J Bradley; Tonnis, Brandon; Pinnow, David; Davis, Jerry; Raymer, Paul; Pederson, Gary A

    2011-09-14

    Castor has tremendous potential as a feedstock for biodiesel production. The oil content and fatty acid composition in castor seed are important factors determining the price for production and affecting the key fuel properties of biodiesel. There are 1033 available castor accessions collected or donated from 48 countries worldwide in the USDA germplasm collection. The entire castor collection was screened for oil content and fatty acid composition by nuclear magnetic resonance (NMR) and gas chromatography (GC), respectively. Castor seeds on the average contain 48.2% oil with significant variability ranging from 37.2 to 60.6%. Methyl esters were prepared from castor seed by alkaline transmethylation. GC analysis of methyl esters confirmed that castor oil was composed primarily of eight fatty acids: 1.48% palmitic (C16:0), 1.58% stearic (C18:0), 4.41% oleic (C18:1), 6.42% linoleic (C18:2), 0.68% linolenic (C18:3), 0.45% gadoleic (C20:1), 84.51% ricinoleic (C18:1-1OH), and 0.47% dihydroxystearic (C18:0-2OH) acids. Significant variability in fatty acid composition was detected among castor accessions. Ricinoleic acid (RA) was positively correlated with dihydroxystearic acid (DHSA) but highly negatively correlated with the five other fatty acids except linolenic acid. The results for oil content and fatty acid composition obtained from this study will be useful for end-users to explore castor germplasm for biodiesel production.

  1. A global review on the influence of beavers (Castor fiber, Castor canadensis) on river and floodplain dynamics

    NASA Astrophysics Data System (ADS)

    Larsen, Annegret; Lane, Stuart; Larsen, Joshua

    2017-04-01

    Beavers (Castor fiber, Castor canadensis) have the ability to actively engineer their habitat, which they can do most effectively in lower order streams and their floodplains. Hence, this engineering has the potential to alter the hydrology, geomorphology, biogeochemistry, and ecology of river systems and the feedbacks between them. Thus, the beaver is often referred to as an 'ecosystem engineer' and is reflected in their recognition as a key species when restoring ecosystems. This capacity to engineer low order streams also shapes a range of positive and negative perceptions on their influence. On the one hand they may be perceived as capable of undermining existing river engineering schemes and the land use of associated floodplains, and on the other hand beavers may provide an alternative to traditional 'hard' engineering, potentially improving river restoration success. The aim of this review is to summarize research to date on the impacts of beavers on stream and floodplain hydrology, geomorphology, water-quality and ecology, and the feedbacks between them. Our review shows that: (1) research has been focused heavily on North American streams, with far less research outside this North American context; (2) there is a tendency to investigate beaver impacts from the perspective of individual disciplines, to the detriment of considering broader process feedbacks, notably at the interface of hydro-geomorphology and riparian ecology; (3) it remains unclear to which extent beavers genuinely engineered streams prior to human impact, pointing to the need for longer term (millennium scale) studies on how beavers have changed river-floodplain systems. Crucially, we conclude that the investigation of the effects of beavers on streams and floodplains, especially in a longer-term, and their use for river restoration can only be understood through the thorough investigation of antecedent hydro-geomorphic conditions which takes account of the ways in which beavers and humans

  2. Castor oil, bath and/or enema for cervical priming and induction of labour.

    PubMed

    Kelly, Anthony J; Kavanagh, Josephine; Thomas, Jane

    2013-07-24

    Castor oil, a potent cathartic, is derived from the bean of the castor plant. Anecdotal reports, which date back to ancient Egypt have suggested the use of castor oil to stimulate labour. Castor oil has been widely used as a traditional method of initiating labour in midwifery practice. Its role in the initiation of labour is poorly understood and data examining its efficacy within a clinical trial are limited. This is one of a series of reviews of methods of cervical ripening and labour induction using standardised methodology. To determine the effects of castor oil or enemas for third trimester cervical ripening or induction of labour in comparison with other methods of cervical ripening or induction of labour. We searched the Cochrane Pregnancy and Childbirth Group's Trials Register (30 April 2013) and bibliographies of relevant papers. Clinical trials comparing castor oil, bath or enemas used for third trimester cervical ripening or labour induction with placebo/no treatment or other methods listed above it on a predefined list of labour induction methods. A strategy was developed to deal with the large volume and complexity of trial data relating to labour induction. This involved a two-stage method of data extraction. Three trials, involving 233 women, are included. There was no evidence of differences in caesarean section rates between the two interventions in the two trials reporting this outcome (risk ratio (RR) 2.04, 95% confidence interval (CI) 0.92 to 4.55). There were no data presented on neonatal or maternal mortality or morbidity.There was no evidence of a difference between castor oil and placebo/no treatment for the rate of instrumental delivery, meconium-stained liquor, or Apgar score less than seven at five minutes. The number of participants was too small to detect all but large differences in outcome. All women who ingested castor oil felt nauseous (RR 59.92, 95% CI 8.46 to 424.52). The three trials included in the review contain small numbers

  3. Monitoring

    SciTech Connect

    Orr, Christopher Henry; Luff, Craig Janson; Dockray, Thomas; Macarthur, Duncan Whittemore

    2004-11-23

    The invention provides apparatus and methods which facilitate movement of an instrument relative to an item or location being monitored and/or the item or location relative to the instrument, whilst successfully excluding extraneous ions from the detection location. Thus, ions generated by emissions from the item or location can successfully be monitored during movement. The technique employs sealing to exclude such ions, for instance, through an electro-field which attracts and discharges the ions prior to their entering the detecting location and/or using a magnetic field configured to repel the ions away from the detecting location.

  4. Identification of diacylglycerol and triacylglycerols containing 11,12,13-trihydroxy-9,14-octadecadienoic acid in castor oil by mass spectrometry

    USDA-ARS?s Scientific Manuscript database

    Castor oil has many industrial uses. Molecular species of acylglycerols containing monohydroxy, dihydroxy and trihydroxy fatty acids in castor oil have been reported. The identification of acylglycerols containing a triOH18:2 fatty acid in castor oil is reported here. The structure of this novel fat...

  5. Identification of diacylglycerol and triacylglycerols containing 11,12,13-trihydroxy-9,14-octadecadienoic acid in castor oil by mass spectrometry

    USDA-ARS?s Scientific Manuscript database

    Castor oil has many industrial uses. Molecular species of acylglycerols containing monohydroxy, dihydroxy and trihydroxy fatty acids in castor oil have been reported. The identification of acylglycerols containing a triOH18:2 fatty acid in castor oil is reported here. The structure of this novel fat...

  6. PICASSO: an end-to-end image simulation tool for space and airborne imaging systems

    NASA Astrophysics Data System (ADS)

    Cota, Stephen A.; Bell, Jabin T.; Boucher, Richard H.; Dutton, Tracy E.; Florio, Christopher J.; Franz, Geoffrey A.; Grycewicz, Thomas J.; Kalman, Linda S.; Keller, Robert A.; Lomheim, Terrence S.; Paulson, Diane B.; Wilkinson, Timothy S.

    2010-06-01

    The design of any modern imaging system is the end result of many trade studies, each seeking to optimize image quality within real world constraints such as cost, schedule and overall risk. Image chain analysis - the prediction of image quality from fundamental design parameters - is an important part of this design process. At The Aerospace Corporation we have been using a variety of image chain analysis tools for many years, the Parameterized Image Chain Analysis & Simulation SOftware (PICASSO) among them. In this paper we describe our PICASSO tool, showing how, starting with a high quality input image and hypothetical design descriptions representative of the current state of the art in commercial imaging satellites, PICASSO can generate standard metrics of image quality in support of the decision processes of designers and program managers alike.

  7. End-to-end information system concept for the Mars Telecommunications Orbiter

    NASA Technical Reports Server (NTRS)

    Breidenthal, Julian C.; Edwards, Charles D.; Greenberg, Edward; Kazz, Greg J.; Noreen, Gary K.

    2006-01-01

    The Mars Telecommunications Orbiter (MTO) was intended to provide high-performance deep space relay links to landers, orbiters, sample-return missions, and approaching spacecraft in the vicinity of Mars, to demostrate interplanetary laser communications, to demonstrate autonomous navigation, and to carry out its own science investigations.

  8. End-to-end information system concept for the Mars Telecommunications Orbiter

    NASA Technical Reports Server (NTRS)

    Bridenthal, Julian C.; Edwards, Charles D.; Greenberg, Edward; Kazz, Greg J.; Noreen, Gary K.

    2006-01-01

    The Mars Telecommunications Orbiter (MTO) was intended to provide high-performance deep space relay links to landers, orbiters, sample-return, missions, and approaching spacecraft in the vicinity of Mars, to demonstrate interplanetary laser communications, to demonstrate autonomous navigation, and to carry out is own science investigations.

  9. Integrating end-to-end encryption and authentication technology into broadband networks

    SciTech Connect

    Pierson, L.G.

    1995-11-01

    BISDN services will involve the integration of high speed data, voice, and video functionality delivered via technology similar to Asynchronous Transfer Mode (ATM) switching and SONET optical transmission systems. Customers of BISDN services may need a variety of data authenticity and privacy assurances, via Asynchronous Transfer Mode (ATM) services Cryptographic methods can be used to assure authenticity and privacy, but are hard to scale for implementation at high speed. The incorporation of these methods into computer networks can severely impact functionality, reliability, and performance. While there are many design issues associated with the serving of public keys for authenticated signaling and for establishment of session cryptovariables, this paper is concerned with the impact of encryption itself on such communications once the signaling and setup have been completed. Network security protections should be carefully matched to the threats against which protection is desired. Even after eliminating unnecessary protections, the remaining customer-required network security protections can impose severe performance penalties. These penalties (further discussed below) usually involve increased communication processing for authentication or encryption, increased error rate, increased communication delay, and decreased reliability/availability. Protection measures involving encryption should be carefully engineered so as to impose the least performance, reliability, and functionality penalties, while achieving the required security protection. To study these trade-offs, a prototype encryptor/decryptor was developed. This effort demonstrated the viability of implementing certain encryption techniques in high speed networks. The research prototype processes ATM cells in a SONET OC-3 payload. This paper describes the functionality, reliability, security, and performance design trade-offs investigated with the prototype.

  10. End-to-End Modeling with the Heimdall Code to Scope High-Power Microwave Systems

    DTIC Science & Technology

    2007-06-01

    which is modeled as a Thevenin-equivalent voltage source, with open-circuit voltage VPP and output impedance ZPP . We model the RF Source electrically...chosen device radius and VPP and ZPP . As a final example, our treatment of the tunable relativistic magnetron uses a piecewise-linear model for...the left and the RF Source to the right, with VPP and ZPP the open-circuit voltage and output impedance of the Pulsed Power and VRF and IRF the

  11. An end-to-end assessment of extreme weather impacts on food security

    NASA Astrophysics Data System (ADS)

    Chavez, Erik; Conway, Gordon; Ghil, Michael; Sadler, Marc

    2015-11-01

    Both governments and the private sector urgently require better estimates of the likely incidence of extreme weather events, their impacts on food crop production and the potential consequent social and economic losses. Current assessments of climate change impacts on agriculture mostly focus on average crop yield vulnerability to climate and adaptation scenarios. Also, although new-generation climate models have improved and there has been an exponential increase in available data, the uncertainties in their projections over years and decades, and at regional and local scale, have not decreased. We need to understand and quantify the non-stationary, annual and decadal climate impacts using simple and communicable risk metrics that will help public and private stakeholders manage the hazards to food security. Here we present an `end-to-end’ methodological construct based on weather indices and machine learning that integrates current understanding of the various interacting systems of climate, crops and the economy to determine short- to long-term risk estimates of crop production loss, in different climate and adaptation scenarios. For provinces north and south of the Yangtze River in China, we have found that risk profiles for crop yields that translate climate into economic variability follow marked regional patterns, shaped by drivers of continental-scale climate. We conclude that to be cost-effective, region-specific policies have to be tailored to optimally combine different categories of risk management instruments.

  12. From End to End: tRNA Editing at 5'- and 3'-Terminal Positions

    PubMed Central

    Betat, Heike; Long, Yicheng; Jackman, Jane E.; Mörl, Mario

    2014-01-01

    During maturation, tRNA molecules undergo a series of individual processing steps, ranging from exo- and endonucleolytic trimming reactions at their 5'- and 3'-ends, specific base modifications and intron removal to the addition of the conserved 3'-terminal CCA sequence. Especially in mitochondria, this plethora of processing steps is completed by various editing events, where base identities at internal positions are changed and/or nucleotides at 5'- and 3'-ends are replaced or incorporated. In this review, we will focus predominantly on the latter reactions, where a growing number of cases indicate that these editing events represent a rather frequent and widespread phenomenon. While the mechanistic basis for 5'- and 3'-end editing differs dramatically, both reactions represent an absolute requirement for generating a functional tRNA. Current in vivo and in vitro model systems support a scenario in which these highly specific maturation reactions might have evolved out of ancient promiscuous RNA polymerization or quality control systems. PMID:25535083

  13. End-to-end observatory software modeling using domain specific languages

    NASA Astrophysics Data System (ADS)

    Filgueira, José M.; Bec, Matthieu; Liu, Ning; Peng, Chien; Soto, José

    2014-07-01

    The Giant Magellan Telescope (GMT) is a 25-meter extremely large telescope that is being built by an international consortium of universities and research institutions. Its software and control system is being developed using a set of Domain Specific Languages (DSL) that supports a model driven development methodology integrated with an Agile management process. This approach promotes the use of standardized models that capture the component architecture of the system, that facilitate the construction of technical specifications in a uniform way, that facilitate communication between developers and domain experts and that provide a framework to ensure the successful integration of the software subsystems developed by the GMT partner institutions.

  14. Improving End-To-End Tsunami Warning for Risk Reduction on Canada’s West Coast

    DTIC Science & Technology

    2015-01-01

    s forest sector’s performance has declined significantly, as have traditional salmon and hering fisheries and seafood processing. Mining activity...transitioning. B.C. has always ofered some of the world’s finest saltwater sport fishing for al five species of Pacific salmon and for botom fish like...Forest Tenure Harvesting Authority • Salmon Hatcheries • Shelfish Aquaculture Plan Unit Boundaries • Saltwater Finfish • Shore unit Polygons

  15. MONTAGE: A Methodology for Designing Composable End-to-End Secure Distributed Systems

    DTIC Science & Technology

    2012-08-01

    of the specification itself. Examples include the seL4 microkernel work by Klein et al. [KEH+09], which presents the experience of formally proving... electromagnetic , optical, acoustic or other nature [AARR02, Kuh02, ST04], or temperature drift [ZBA10]. We make the assumption that virtual machines have no

  16. End-to-end design consideration of a radar altimeter for terrain-aided navigation

    NASA Astrophysics Data System (ADS)

    Chun, Joohwan; Choi, Sanghyouk; Paek, Inchan; Park, Dongmin; Yoo, Kyungju

    2013-10-01

    We present a preliminary simulation study of an interferometric SAR altimeter for the terrain-aided navigation application. Our simulation includes raw SAR data generation, azimuth compression, leading edge detection of the echo signal, maximum likelihood angle estimation and the Bayesian state estimation. Sour results show that radar altimeter performance can be improved with the feedback loop from the rear-end navigation part.

  17. Post2 End-to-End Descent and Landing Simulation for ALHAT Design Analysis Cycle 2

    NASA Technical Reports Server (NTRS)

    Davis, Jody L.; Striepe, Scott A.; Maddock, Robert W.; Johnson, Andrew E.; Paschall, Stephen C., II

    2010-01-01

    The ALHAT project is an agency-level program involving NASA centers, academia, and industry, with a primary goal to develop a safe, autonomous, precision-landing system for robotic and crew-piloted lunar and planetary descent vehicles. POST2 is used as the 6DOF descent and landing trajectory simulation for determining integrated system performance of ALHAT landing-system models and lunar environment models. This paper presents updates in the development of the ALHAT POST2 simulation, as well as preliminary system performance analysis for ALDAC-2 used for the testing and assessment of ALHAT system models. The ALDAC-2 POST2 Monte Carlo simulation results have been generated and focus on HRN model performance with the fully integrated system, as well performance improvements of AGNC and TSAR model since the previous design analysis cycle

  18. Australasian brachytherapy audit: results of the 'end-to-end' dosimetry pilot study.

    PubMed

    Haworth, Annette; Wilfert, Lisa; Butler, Duncan; Ebert, Martin A; Todd, Stephen; Bucci, Joseph; Duchesne, Gillian M; Joseph, David; Kron, Tomas

    2013-08-01

    We present the results of a pilot study to test the feasibility of a brachytherapy dosimetry audit. The feasibility study was conducted at seven sites from four Australian states in both public and private centres. A purpose-built cylindrical water phantom was imaged using the local imaging protocol and a treatment plan was generated to deliver 1 Gy to the central (1 of 3) thermoluminescent dosimeter (TLD) from six dwell positions. All centres completed the audit, consisting of three consecutive irradiations, within a 2-h time period, with the exception of one centre that uses a pulsed dose rate brachytherapy unit. All TLD results were within 4.5% of the predicted value, with the exception of one subset where the dwell position step size was incorrectly applied. While the limited data collected in the study demonstrated considerable heterogeneity in clinical practice, the study proved a brachytherapy dosimetry audit to be feasible. Future studies should include verification of source strength using a Standard Dosimetry Laboratory calibrated chamber, a phantom that more closely mimics the clinical situation, a more comprehensive review of safety and quality assurance (QA) procedures including source dwell time and position accuracy, and a review of patient treatment QA procedures such as applicator position verification. © 2013 The Authors. Journal of Medical Imaging and Radiation Oncology © 2013 The Royal Australian and New Zealand College of Radiologists.

  19. Comparison of Reconstruction and Control algorithms on the ESO end-to-end simulator OCTOPUS

    NASA Astrophysics Data System (ADS)

    Montilla, I.; Béchet, C.; Lelouarn, M.; Correia, C.; Tallon, M.; Reyes, M.; Thiébaut, É.

    Extremely Large Telescopes are very challenging concerning their Adaptive Optics requirements. Their diameters, the specifications demanded by the science for which they are being designed for, and the planned use of Extreme Adaptive Optics systems, imply a huge increment in the number of degrees of freedom in the deformable mirrors. It is necessary to study new reconstruction algorithms to implement the real time control in Adaptive Optics at the required speed. We have studied the performance, applied to the case of the European ELT, of three different algorithms: the matrix-vector multiplication (MVM) algorithm, considered as a reference; the Fractal Iterative Method (FrIM); and the Fourier Transform Reconstructor (FTR). The algorithms have been tested on ESO's OCTOPUS software, which simulates the atmosphere, the deformable mirror, the sensor and the closed-loop control. The MVM is the default reconstruction and control method implemented in OCTOPUS, but it scales in O(N2) operations per loop so it is not considered as a fast algorithm for wave-front reconstruction and control on an Extremely Large Telescope. The two other methods are the fast algorithms studied in the E-ELT Design Study. The performance, as well as their response in the presence of noise and with various atmospheric conditions, has been compared using a Single Conjugate Adaptive Optics configuration for a 42 m diameter ELT, with a total amount of 5402 actuators. Those comparisons made on a common simulator allow to enhance the pros and cons of the various methods, and give us a better understanding of the type of reconstruction algorithm that an ELT demands.

  20. End-to-end laser radar range code for coherent cw lasers

    NASA Astrophysics Data System (ADS)

    Yoder, M. John; Seliverstov, Dima

    1996-06-01

    A user friendly modular computer code is described for CW coherent laser radar which includes all relevant physical effects needed to evaluate the probability of detection versus time after launch for ballistic missiles or other targets of interest. The beginning point of the code is the conventional laser radar range equation. Atmospheric attenuation is determined from an integral FASCODE calculation, and the laser radar range equation is solved for a curved-earth geometry including free air turbulence induced beam spreading. Several different atmospheric turbulence models are selectable. Target cross-sections can be input into the code as a function of aspect angle Coherence time and transverse coherence length limits are included in the code. Beam jitter effects are also calculated. The carrier-to-noise ratio is calculated including all of these (complicated) variables and degradations. The code then calculates the probability of detection of the target as a function of time using incoherent integration of coherent sub-pulses. The governing equations and practical results are presented for detection and tracking of long range theater ballistic missiles from airborne surveillance platforms. The use of CW lasers requires increased measurement times compared to pulsed lasers and results in an averaging of the target fading statistics.

  1. Assessing the Performance Limits of Internal Coronagraphs Through End-to-End Modeling

    NASA Technical Reports Server (NTRS)

    Krist, John E.; Belikov, Ruslan; Pueyo, Laurent; Mawet, Dimitri P.; Moody, Dwight; Trauger, John T.; Shaklan, Stuart B.

    2013-01-01

    As part of the NASA ROSES Technology Demonstrations for Exoplanet Missions (TDEM) program, we conducted a numerical modeling study of three internal coronagraphs (PIAA, vector vortex, hybrid bandlimited) to understand their behaviors in realistically-aberrated systems with wavefront control (deformable mirrors). This investigation consisted of two milestones: (1) develop wavefront propagation codes appropriate for each coronagraph that are accurate to 1% or better (compared to a reference algorithm) but are also time and memory efficient, and (2) use these codes to determine the wavefront control limits of each architecture. We discuss here how the milestones were met and identify some of the behaviors particular to each coronagraph. The codes developed in this study are being made available for community use. We discuss here results for the HBLC and VVC systems, with PIAA having been discussed in a previous proceeding.

  2. Assessing the Performance Limits of Internal Coronagraphs Through End-to-End Modeling

    NASA Technical Reports Server (NTRS)

    Krist, John E.; Belikov, Ruslan; Pueyo, Laurent; Mawet, Dimitri P.; Moody, Dwight; Trauger, John T.; Shaklan, Stuart B.

    2013-01-01

    As part of the NASA ROSES Technology Demonstrations for Exoplanet Missions (TDEM) program, we conducted a numerical modeling study of three internal coronagraphs (PIAA, vector vortex, hybrid bandlimited) to understand their behaviors in realistically-aberrated systems with wavefront control (deformable mirrors). This investigation consisted of two milestones: (1) develop wavefront propagation codes appropriate for each coronagraph that are accurate to 1% or better (compared to a reference algorithm) but are also time and memory efficient, and (2) use these codes to determine the wavefront control limits of each architecture. We discuss here how the milestones were met and identify some of the behaviors particular to each coronagraph. The codes developed in this study are being made available for community use. We discuss here results for the HBLC and VVC systems, with PIAA having been discussed in a previous proceeding.

  3. HIDE & SEEK: End-to-end packages to simulate and process radio survey data

    NASA Astrophysics Data System (ADS)

    Akeret, J.; Seehars, S.; Chang, C.; Monstein, C.; Amara, A.; Refregier, A.

    2017-01-01

    As several large single-dish radio surveys begin operation within the coming decade, a wealth of radio data will become available and provide a new window to the Universe. In order to fully exploit the potential of these datasets, it is important to understand the systematic effects associated with the instrument and the analysis pipeline. A common approach to tackle this is to forward-model the entire system-from the hardware to the analysis of the data products. For this purpose, we introduce two newly developed, open-source Python packages: the HI Data Emulator (HIDE) and the Signal Extraction and Emission Kartographer (SEEK) for simulating and processing single-dish radio survey data. HIDE forward-models the process of collecting astronomical radio signals in a single-dish radio telescope instrument and outputs pixel-level time-ordered-data. SEEK processes the time-ordered-data, removes artifacts from Radio Frequency Interference (RFI), automatically applies flux calibration, and aims to recover the astronomical radio signal. The two packages can be used separately or together depending on the application. Their modular and flexible nature allows easy adaptation to other instruments and datasets. We describe the basic architecture of the two packages and examine in detail the noise and RFI modeling in HIDE, as well as the implementation of gain calibration and RFI mitigation in SEEK. We then apply HIDE &SEEK to forward-model a Galactic survey in the frequency range 990-1260 MHz based on data taken at the Bleien Observatory. For this survey, we expect to cover 70% of the full sky and achieve a median signal-to-noise ratio of approximately 5-6 in the cleanest channels including systematic uncertainties. However, we also point out the potential challenges of high RFI contamination and baseline removal when examining the early data from the Bleien Observatory. The fully documented HIDE &SEEK packages are available at http://hideseek.phys.ethz.ch/ and are published under the GPLv3 license on GitHub.

  4. Mars End-to-End Microfluidic Analyzer (MEEMA) for Solids, Liquids, and Gases

    NASA Astrophysics Data System (ADS)

    Willis, P. A.; Stockton, A. M.; Mora, M. F.; Cable, M. L.; Jensen, E. C.; Jiao, H.; Mathies, R. A.

    2012-06-01

    We propose to develop MEEMA, a 3kg, 2W, flight-capable microfluidic lab-on-a-chip analyzer that is capable of injesting solid, liquid, or gas samples and performing a suite of chemical analyses with parts per trillion sensitivity.

  5. Standardization of End-to-End Performance of Digital Video Teleconferencing/Video Telephony Systems

    DTIC Science & Technology

    1991-12-01

    Teleconferencing, NCS-TIB-90-6, Temporal Frequency as a Technique to Measure the Ability of a Teleconferencing System to Reproduce Motion, NCS-TIB-91...used for most tests. For completeness and ease of reference, the basic parameters of all patterns are repeated on Table 3-3, while a sample of one...video signal which influences the signal-to-noise ratio but does not have primary impact on the ability of the codec to faithfully reproduce a moving

  6. Hardware Support for Malware Defense and End-to-End Trust

    DTIC Science & Technology

    2017-02-01

    10 Figure 4 Design flow of ACM Bluespec FPGA model ................................................................ 11 Figure 5 ACM concept...may be insufficient for continuation of trust at an arbitrary point in the future. If the design of the system permits undetectable attacks against...formed IoT division, and with selected partners.  Standards: “Trust Dust”, the early prototype of the value of embedded TPM for cyber- physical systems

  7. Topological Constraints on Identifying Additive Link Metrics via End-to-end Paths Measurements

    DTIC Science & Technology

    2012-09-20

    U.K. Ministry of Defence or the U.K. Government. The U.S. and U.K. Governments are authorized to reproduce and distribute reprints for Government...MONITOR’S REPORT NUMBER(S) 12. DISTRIBUTION /AVAILABILITY STATEMENT Approved for public release; distribution unlimited 13. SUPPLEMENTARY NOTES...with (partially) unknown probabil- ity distributions , and apply various parametric/nonparametric techniques to estimate the link metric distributions

  8. PICASSO: an end-to-end image simulation tool for space and airborne imaging systems

    NASA Astrophysics Data System (ADS)

    Cota, Steve A.; Bell, Jabin T.; Boucher, Richard H.; Dutton, Tracy E.; Florio, Chris J.; Franz, Geoffrey A.; Grycewicz, Thomas J.; Kalman, Linda S.; Keller, Robert A.; Lomheim, Terrence S.; Paulson, Diane B.; Willkinson, Timothy S.

    2008-08-01

    The design of any modern imaging system is the end result of many trade studies, each seeking to optimize image quality within real world constraints such as cost, schedule and overall risk. Image chain analysis - the prediction of image quality from fundamental design parameters - is an important part of this design process. At The Aerospace Corporation we have been using a variety of image chain analysis tools for many years, the Parameterized Image Chain Analysis & Simulation SOftware (PICASSO) among them. In this paper we describe our PICASSO tool, showing how, starting with a high quality input image and hypothetical design descriptions representative of the current state of the art in commercial imaging satellites, PICASSO can generate standard metrics of image quality in support of the decision processes of designers and program managers alike.

  9. Privacy in Pharmacogenetics: An End-to-End Case Study of Personalized Warfarin Dosing

    PubMed Central

    Fredrikson, Matthew; Lantz, Eric; Jha, Somesh; Lin, Simon; Page, David; Ristenpart, Thomas

    2014-01-01

    We initiate the study of privacy in pharmacogenetics, wherein machine learning models are used to guide medical treatments based on a patient’s genotype and background. Performing an in-depth case study on privacy in personalized warfarin dosing, we show that suggested models carry privacy risks, in particular because attackers can perform what we call model inversion: an attacker, given the model and some demographic information about a patient, can predict the patient’s genetic markers. As differential privacy (DP) is an oft-proposed solution for medical settings such as this, we evaluate its effectiveness for building private versions of pharmacogenetic models. We show that DP mechanisms prevent our model inversion attacks when the privacy budget is carefully selected. We go on to analyze the impact on utility by performing simulated clinical trials with DP dosing models. We find that for privacy budgets effective at preventing attacks, patients would be exposed to increased risk of stroke, bleeding events, and mortality. We conclude that current DP mechanisms do not simultaneously improve genomic privacy while retaining desirable clinical efficacy, highlighting the need for new mechanisms that should be evaluated in situ using the general methodology introduced by our work. PMID:27077138

  10. End-to-end simulation of the image stability for the airborne telescope SOFIA

    NASA Astrophysics Data System (ADS)

    Schoenhoff, Ulrich; Eisentraeger, Peter; Wandner, Karl; Kaercher, Hans J.; Nordmann, Rainer

    2000-06-01

    To provide astronomers access to infrared wavelength unavailable from the ground the airborne telescope SOFIA is in development. This paper focuses on the image stability of the telescope, its modeling and simulation. The operation of the telescope under the harsh environmental conditions in the aircraft makes the prediction of the image stability during the design process necessary. For this purpose an integrated mathematical simulation model, which includes the optics, the structural dynamics and the control loops has been constructed. Because of the high relevance of the structural dynamics for image stability and control design, special attention is paid to the import and reduction of the finite element model of the telescopes mechanical structure. Different control approaches are considered for the attitude control and the compensation of the impact of the structural flexibility on the image motion. Additionally the secondary mirror servo-mechanism is utilized to optimize the image stability. Simulation results are shown.

  11. Design and Evaluation for the End-to-End Detection of TCP/IP Header Manipulation

    DTIC Science & Technology

    2014-06-01

    systems making IPv4 and IPv6 address announcements (over 56 thousand as of May 2014 [90]) and the number of domains observed in DNS (over 168 million...if they change an immutable field. Another point to note is that IPv6 , the anticipated eventual replacement for IPv4, does not even include a...protocols such as UDP and IPv6 may help improve the inferences HICCUPS provides to TCP or help a system’s networking stack improve performance for

  12. EQUIP: end-to-end quantification of uncertainty for impacts prediction

    NASA Astrophysics Data System (ADS)

    Morse, A. P.; Challinor, A. J.; Equip Consortium

    2010-12-01

    Inherent uncertainties in climate prediction present a serious challenge to attempts to assess future impacts and adaptation options. Such assessments are critical to any policy decisions regarding investment in resources to ensure human and environmental wellbeing in the face of environmental change and a growing population. Current methods for quantifying uncertainty in projections of climate and its impacts tend to focus first on taking full account of uncertainty, with a subsequent step assessing utility. We argue that a new approach is required, whereby climate and impacts models are used to develop risk-based prediction systems that focus on the information content of models and utility for decision-making. Preliminary steps in this direction are explored, principally using the example of climate-induced changes in crop yield. The relative contribution of uncertainty in crop and climate simulation to the total uncertainty in projected yield changes is examined. A focus on governing bio-physical processes across a number of crop models is used to characterise the robustness of the results. Further development of this approach relies on the development of decision-focussed techniques that analyse sources of uncertainty and assess and improve the information content of models of climate and its impacts. Such an approach is significantly different from tagging impacts models onto climate models. It implies substantial interaction with other organisations and stakeholders from development NGOs to the insurance sector and policy makers. These interactions should be aimed at ensuring that the principal lead-times, and formats, for the impact projections are those relevant to decision-making. The EQUIP project, and its associated open network of scientists, aims to develop the approach outlined above. The project is examining the cascade of uncertainty from climate to impacts by conducting integrated analyses of a range of sectors, principally crops, marine ecosystems, water management, heat waves and droughts. The research includes assessment of the information content of climate model projections, combination of climate models and data-driven models to support decisions, and evaluation of the quality of climate and impacts predictions.

  13. Emergence of Laplace therapeutics: declaring an end to end-stage heart failure.

    PubMed

    Mehra, Mandeep R; Uber, Patricia A

    2002-01-01

    A large number of chronic heart failure patients escape from the benefits of neurohormonal blockade only to transit into a discouragingly miserable state of what the physician often refers to as end-stage heart failure. Conceptually, the designation of end-stage as a description of a clinical scenario implies pessimism concerning recourse to a therapeutic avenue. A variety of surgical therapeutic techniques that take advantage of the law of Laplace, designed to effectively restore the cardiac shape from a spherical, mechanically inefficient pump to a more elliptical, structurally sound organ are now being employed. Additionally, the field of mechanical device implantation is surging ahead at a rapid pace. The weight of evidence regarding mechanical unloading using assist devices suggests that hemodynamic restoration is accompanied by regression of cellular hypertrophy, normalization of the neuroendocrine axis, improved expression of contractile proteins, enhanced cellular respiratory control, and decreases in markers of apoptosis and cellular stress. Thus, these lines of data point toward discarding the notion of end-stage heart failure. We are at a new crossroad in our quest to tackle chronic heart failure. It is our contention that the use of antiremodeling strategies, including device approaches, will soon signal the end of end-stage heart failure.

  14. End-to-End Verification of Information-Flow Security for C and Assembly Programs

    DTIC Science & Technology

    2016-04-01

    sentation purposes, we omit details regarding the well-known “infinite stuttering ” problem for simulations (described, for example, in [14]). Our Coq...definition of simulation includes a well-founded order that prevents infinite stuttering . Also notice that, contrary to our discussion earlier, we do

  15. Non-adaptive End-to-End Diagnosis for Complex Networks

    DTIC Science & Technology

    2015-09-30

    CISS, 2014. [5] M. Wang, W. Xu, E. Mallada and A. Tang, “Sparse Recovery with Graph Constraints”, IEEE Transactions on Information Theory ...of number of measurements for any given graph . In particular, for several special graphs , we were able to find the exact optimal number of...for goal number 1, we have shown the lower bound of number of measurements for any given graph . In particular, for several special graphs , we were

  16. Experiments with Memory-to-Memory Coupling for End-to-End fusion Simulation Workflows

    SciTech Connect

    Docan, Ciprian; Zhang, Fan; Parashar, Manish; Cummings, Julian; Podhorszki, Norbert; Klasky, Scott A

    2010-01-01

    Scientific applications are striving to accurately simulate multiple interacting physical processes that comprise complex phenomena being modeled. Efficient and scalable parallel implementations of these coupled simulations present challenging interaction and coordination requirements, especially when the coupled physical processes are computationally heterogeneous and progress at different speeds. In this paper, we present the design, implementation and evaluation of a memory-to-memory coupling framework for coupled scientific simulations on high-performance parallel computing platforms. The framework is driven by the coupling requirements of the Center for Plasma Edge Simulation, and it provides simple coupling abstractions as well as efficient asynchronous (RDMA-based) memory-to-memory data transport mechanisms that complement existing parallel programming systems and data sharing frameworks. The framework enables flexible coupling behaviors that are asynchronous in time and space, and it supports dynamic coupling between heterogeneous simulation processes without enforcing any synchronization constraints. We evaluate the performance and scalability of the coupling framework using a specific coupling scenario, on the Jaguar Cray XT5 system at Oak Ridge National Laboratory.

  17. Increasing Army Supply Chain Performance: Using an Integrated End to End Metrics System

    DTIC Science & Technology

    2017-01-01

    performance specifications and feedback loops for ongoing and dynamic performance measurement (Bititci, Carrie, & McDevitt, 1997; Melnyk, Stewart ... Stewart , D. M., Zsidisin, G. A., Hanson, J., & Burns, L. (2005). An empirical investigation of the metrics alignment process. International Journal...of Productivity and Performance Management, 54(5/6), 312–324. Melnyk, S. A., Stewart , D. M., & Swink, M. (2004). Metrics and performance measurement

  18. Stock assessment and end-to-end ecosystem models alter dynamics of fisheries data.

    PubMed

    Storch, Laura S; Glaser, Sarah M; Ye, Hao; Rosenberg, Andrew A

    2017-01-01

    Although all models are simplified approximations of reality, they remain useful tools for understanding, predicting, and managing populations and ecosystems. However, a model's utility is contingent on its suitability for a given task. Here, we examine two model types: single-species fishery stock assessment and multispecies marine ecosystem models. Both are efforts to predict trajectories of populations and ecosystems to inform fisheries management and conceptual understanding. However, many of these ecosystems exhibit nonlinear dynamics, which may not be represented in the models. As a result, model outputs may underestimate variability and overestimate stability. Using nonlinear forecasting methods, we compare predictability and nonlinearity of model outputs against model inputs using data and models for the California Current System. Compared with model inputs, time series of model-processed outputs show more predictability but a higher prevalence of linearity, suggesting that the models misrepresent the actual predictability of the modeled systems. Thus, caution is warranted: using such models for management or scenario exploration may produce unforeseen consequences, especially in the context of unknown future impacts.

  19. Modeling and Simulation of Satellite Subsystems for End-to-End Spacecraft Modeling

    DTIC Science & Technology

    2006-04-01

    reaction wheel power requirements for a particular maneuver. The ADCS only calculates information at a specified time and passes it to SST; no future or...are calculated and fed into a reaction wheel model which will return the power requirements for the maneuver. In the case of calculating slew time

  20. An end-to-end workflow for engineering of biological networks from high-level specifications.

    PubMed

    Beal, Jacob; Weiss, Ron; Densmore, Douglas; Adler, Aaron; Appleton, Evan; Babb, Jonathan; Bhatia, Swapnil; Davidsohn, Noah; Haddock, Traci; Loyall, Joseph; Schantz, Richard; Vasilev, Viktor; Yaman, Fusun

    2012-08-17

    We present a workflow for the design and production of biological networks from high-level program specifications. The workflow is based on a sequence of intermediate models that incrementally translate high-level specifications into DNA samples that implement them. We identify algorithms for translating between adjacent models and implement them as a set of software tools, organized into a four-stage toolchain: Specification, Compilation, Part Assignment, and Assembly. The specification stage begins with a Boolean logic computation specified in the Proto programming language. The compilation stage uses a library of network motifs and cellular platforms, also specified in Proto, to transform the program into an optimized Abstract Genetic Regulatory Network (AGRN) that implements the programmed behavior. The part assignment stage assigns DNA parts to the AGRN, drawing the parts from a database for the target cellular platform, to create a DNA sequence implementing the AGRN. Finally, the assembly stage computes an optimized assembly plan to create the DNA sequence from available part samples, yielding a protocol for producing a sample of engineered plasmids with robotics assistance. Our workflow is the first to automate the production of biological networks from a high-level program specification. Furthermore, the workflow's modular design allows the same program to be realized on different cellular platforms simply by swapping workflow configurations. We validated our workflow by specifying a small-molecule sensor-reporter program and verifying the resulting plasmids in both HEK 293 mammalian cells and in E. coli bacterial cells.