Sample records for haystack large-scale identification

  1. Cost Recovery: Finding a Needle in the Haystack.

    ERIC Educational Resources Information Center

    Colucci, Michael A.

    1988-01-01

    The first step in cost recovery for asbestos removal is identification of product and manufacturer. Tables illustrate sources of product identification, types of asbestos bulk sample reports and source documents, asbestos-related diseases, and the synergistic effect of asbestos and cigarette smoking. (MLF)

  2. Final Report of the Haystack Orbital Debris Data Review Panel

    NASA Technical Reports Server (NTRS)

    Barton, David K.; Brillinger, David; McDaniel, Patrick; Pollock, Kenneth H.; El-Shaarawi, A. H.; Tuley, Michael T.

    1998-01-01

    The Haystack Orbital Debris Data Review Panel was established in December 1996 to consider the adequacy of the data on orbital debris gathered over the past several years with the Haystack radar, and the accuracy of the methods used to estimate the flux vs. size relationship for this debris. The four specific issues addressed for the Panel were: 1. The number of observations relative to the estimated population of interest 2. The inherent ambiguity between the measured radar cross section (RCS) and the inferred physical size of the object 3. The inherent aspect angle limitation in viewing each object and its relationship to object geometry 4. The adequacy of the sample data set to characterize the debris population's potential geometry. Further discussion and interpretation of these issues, and identification of the detailed questions contributing to them, are discussed in this report.

  3. Finding Needles in Haystacks: Identity Mismatch Frequency and Facial Identity Verification

    ERIC Educational Resources Information Center

    Bindemann, Markus; Avetisyan, Meri; Blackwell, Kristy-Ann

    2010-01-01

    Accurate person identification is central to all security, police, and judicial systems. A commonplace method to achieve this is to compare a photo-ID and the face of its purported owner. The critical aspect of this task is to spot cases in which these two instances of a face do not match. Studies of person identification show that these instances…

  4. DOE Office of Scientific and Technical Information (OSTI.GOV)

    The open source Project Haystack initiative defines meta data and communication standards related to data from buildings and intelligent devices. The Project Haystack REST API defines standard formats and operations for exchanging Haystack tagged data over HTTP. The HaystackRuby gem wraps calls to this REST API to enable Ruby application to easily integrate data hosted on a Project Haystack compliant server. The HaystackRuby gem was developed at the National Renewable Energy Lab to support applications related to campus energy. We hope that this tool may be useful to others.

  5. The British in Kenya (1952-1960): Analysis of a Successful Counterinsurgency Campaign

    DTIC Science & Technology

    2005-06-01

    East Africa to seek their riches in cattle , coffee, mining, and selling safaris to tourists. While the other colonial powers set their sights on...violence on a large scale by early 1952. Cattle on white settlements were being destroyed and standing crops and haystacks set on fire, particularly...Guards used pliers to castrate Mau Mau prisoners. Whatever the method and level of brutality, by the latter half of the 1950s most Kikuyu had turned

  6. Haystack, a web-based tool for metabolomics research

    PubMed Central

    2014-01-01

    Background Liquid chromatography coupled to mass spectrometry (LCMS) has become a widely used technique in metabolomics research for differential profiling, the broad screening of biomolecular constituents across multiple samples to diagnose phenotypic differences and elucidate relevant features. However, a significant limitation in LCMS-based metabolomics is the high-throughput data processing required for robust statistical analysis and data modeling for large numbers of samples with hundreds of unique chemical species. Results To address this problem, we developed Haystack, a web-based tool designed to visualize, parse, filter, and extract significant features from LCMS datasets rapidly and efficiently. Haystack runs in a browser environment with an intuitive graphical user interface that provides both display and data processing options. Total ion chromatograms (TICs) and base peak chromatograms (BPCs) are automatically displayed, along with time-resolved mass spectra and extracted ion chromatograms (EICs) over any mass range. Output files in the common .csv format can be saved for further statistical analysis or customized graphing. Haystack's core function is a flexible binning procedure that converts the mass dimension of the chromatogram into a set of interval variables that can uniquely identify a sample. Binned mass data can be analyzed by exploratory methods such as principal component analysis (PCA) to model class assignment and identify discriminatory features. The validity of this approach is demonstrated by comparison of a dataset from plants grown at two light conditions with manual and automated peak detection methods. Haystack successfully predicted class assignment based on PCA and cluster analysis, and identified discriminatory features based on analysis of EICs of significant bins. Conclusion Haystack, a new online tool for rapid processing and analysis of LCMS-based metabolomics data is described. It offers users a range of data visualization options and supports non-biased differential profiling studies through a unique and flexible binning function that provides an alternative to conventional peak deconvolution analysis methods. PMID:25350247

  7. Haystack, a web-based tool for metabolomics research.

    PubMed

    Grace, Stephen C; Embry, Stephen; Luo, Heng

    2014-01-01

    Liquid chromatography coupled to mass spectrometry (LCMS) has become a widely used technique in metabolomics research for differential profiling, the broad screening of biomolecular constituents across multiple samples to diagnose phenotypic differences and elucidate relevant features. However, a significant limitation in LCMS-based metabolomics is the high-throughput data processing required for robust statistical analysis and data modeling for large numbers of samples with hundreds of unique chemical species. To address this problem, we developed Haystack, a web-based tool designed to visualize, parse, filter, and extract significant features from LCMS datasets rapidly and efficiently. Haystack runs in a browser environment with an intuitive graphical user interface that provides both display and data processing options. Total ion chromatograms (TICs) and base peak chromatograms (BPCs) are automatically displayed, along with time-resolved mass spectra and extracted ion chromatograms (EICs) over any mass range. Output files in the common .csv format can be saved for further statistical analysis or customized graphing. Haystack's core function is a flexible binning procedure that converts the mass dimension of the chromatogram into a set of interval variables that can uniquely identify a sample. Binned mass data can be analyzed by exploratory methods such as principal component analysis (PCA) to model class assignment and identify discriminatory features. The validity of this approach is demonstrated by comparison of a dataset from plants grown at two light conditions with manual and automated peak detection methods. Haystack successfully predicted class assignment based on PCA and cluster analysis, and identified discriminatory features based on analysis of EICs of significant bins. Haystack, a new online tool for rapid processing and analysis of LCMS-based metabolomics data is described. It offers users a range of data visualization options and supports non-biased differential profiling studies through a unique and flexible binning function that provides an alternative to conventional peak deconvolution analysis methods.

  8. Identification of a Debris Cloud from the Nuclear Powered SNAPSHOT Satellite with Haystack Radar Measurements

    NASA Technical Reports Server (NTRS)

    Stokely, C.; Stansbery, E.

    2006-01-01

    Data from the MIT Lincoln Laboratory (MIT/LL) Long Range Imaging Radar (known as the Haystack radar) have been used in the past to examine families of objects from individual satellite breakups or families of orbiting objects that can be isolated in altitude and inclination. This is possible because for some time after a breakup, the debris cloud of particles can remain grouped together in similar orbit planes. This cloud will be visible to the radar, in fixed staring mode, for a short time twice each day, as the orbit plane moves through the field of view. There should be a unique three-dimensional pattern in observation time, range, and range rate which can identify the cloud. Eventually, through slightly differing precession rates of the right ascension of ascending node of the debris cloud, the observation time becomes distributed so that event identification becomes much more difficult. Analyses of the patterns in observation time, range, and range rate have identified good debris candidates released from the polar orbiting SNAPSHOT satellite (International Identifier: 1965-027A). For orbits near 90o inclination, there is essentially no precession of the orbit plane. The SNAPSHOT satellite is a well known nuclear powered satellite launched in 1965 to a near circular 1300 km orbit with an inclination of 90.3o. This satellite began releasing debris in 1979 with new pieces being discovered and cataloged over the years. 51 objects are still being tracked by the United States Space Surveillance Network. An analysis of the Haystack data has identified at least 60 pieces of debris separate from the 51 known tracked debris pieces, where all but 2 of the 60 pieces have a size less than 10cm. The altitude and inclination (derived from range-rate with a circular orbit assumption) are consistent with the SNAPSHOT satellite and its tracked debris cloud.

  9. Modeling of the Orbital Debris Population of RORSAT Sodium-Potassium Droplets

    NASA Technical Reports Server (NTRS)

    Xu, Y.-L.; Krisko, P. H.; Matney, Mark; Stansbery, E. G.

    2010-01-01

    A large population resident in the orbital debris environment is composed of eutectic sodium-potassium (NaK) droplets, released during the reactor core ejection of 16 nuclear-powered Radar Ocean Reconnaissance Satellites (RORSATs) launched in the 1980s by the former Soviet Union. These electrically conducting RORSAT debris objects are spherical in shape, generating highly polarized radar returns. Their diameters are mostly in the centimeter and millimeter size regimes. Since the Space Surveillance Network catalog is limited to objects greater than 5 cm in low Earth orbit, our current knowledge about this special class of orbital debris relies largely on the analysis of Haystack radar data. This paper elaborates the simulation of the RORSAT debris populations in the new NASA Orbital Debris Engineering Model ORDEM2010, which replaces ORDEM2000. The estimation of the NaK populations uses the NASA NaK-module as a benchmark. It follows the general statistical approach to developing all other ORDEM2010-required LEO populations (for various types of debris and across a wide range of object sizes). This paper describes, in detail, each major step in the NaK-population derivation, including a specific discussion on the conversion between Haystack-measured radar-cross-sections and object-size distribution for the NaK droplets. Modeling results show that the RORSAT debris population is stable for the time period under study and that Haystack data sets are fairly consistent over the observations of multiple years.

  10. Pinpointing Needles in Giant Haystacks: Use of Text Mining to Reduce Impractical Screening Workload in Extremely Large Scoping Reviews

    ERIC Educational Resources Information Center

    Shemilt, Ian; Simon, Antonia; Hollands, Gareth J.; Marteau, Theresa M.; Ogilvie, David; O'Mara-Eves, Alison; Kelly, Michael P.; Thomas, James

    2014-01-01

    In scoping reviews, boundaries of relevant evidence may be initially fuzzy, with refined conceptual understanding of interventions and their proposed mechanisms of action an intended output of the scoping process rather than its starting point. Electronic searches are therefore sensitive, often retrieving very large record sets that are…

  11. Haystack Observatory Technology Development Center

    NASA Technical Reports Server (NTRS)

    Beaudoin, Chris; Corey, Brian; Niell, Arthur; Cappallo, Roger; Whitney, Alan

    2013-01-01

    Technology development at MIT Haystack Observatory were focused on four areas in 2012: VGOS developments at GGAO; Digital backend developments and workshop; RFI compatibility at VLBI stations; Mark 6 VLBI data system development.

  12. Ionospheric Modeling: Development, Verification and Validation

    DTIC Science & Technology

    2005-09-01

    facilitate the automated processing of a large network of GPS receiver data. 4.; CALIBRATION AND VALIDATION OF IONOSPHERIC SENSORS We have been...NOFS Workshop, Estes Park, CO, January 2005. W. Rideout, A. Coster, P. Doherty, MIT Haystack Automated Processing of GPS Data to Produce Worldwide TEC

  13. Project Haystack: The Search for Life in the Galaxy.

    ERIC Educational Resources Information Center

    Search for Extraterrestrial Intelligence Inst., Mountain View, CA.

    Produced by the Search for Extraterrestrial Intelligence (SETI), Project Haystack presents scenarios that depict various aspects of the search for extraterrestrial intelligence. Students conduct hands-on and minds-on activities while exploring what it means to send and receive a message across interstellar distances. Students explore and map vast…

  14. Detecting bacteria in food- harder than searching for a needle in a haystack

    USDA-ARS?s Scientific Manuscript database

    Actually, I think it would have been easier to find a needle in a haystack than to locate variables pertaining to that time-honored question and mathematically compute the relative difficulty of detecting bacteria versus searching for aforementioned needle. Perhaps it might be unusual given that I ...

  15. Haystack Analysis Center

    NASA Technical Reports Server (NTRS)

    Niell, Arthur; Cappallo, Roger; Corey, Brian; Titus, Mike

    2013-01-01

    Analysis activities at Haystack Observatory are directed towards improving the accuracy of geodetic measurements, whether these are from VLBI, GNSS, SLR, or any other technique. Those analysis activities that are related to technology development are reported elsewhere in this volume. In this report, a preliminary analysis of the first geodetic sessions with the new broadband geodetic VLBI system is reported.

  16. Space Debris Symposium (A6.) Measurements and Space Surveillance (1.): Measurements of the Small Particle Debris Cloud from the 11 January, 2007 Chinese Anti-satellite Test

    NASA Technical Reports Server (NTRS)

    Matney, Mark J.; Stansbery, Eugene; J.-C Liou; Stokely, Christopher; Horstman, Matthew; Whitlock, David

    2008-01-01

    On January 11, 2007, the Chinese military conducted a test of an anti-satellite (ASAT) system, destroying their own Fengyun-1C spacecraft with an interceptor missile. The resulting hypervelocity collision created an unprecedented number of tracked debris - more than 2500 objects. These objects represent only those large enough for the US Space Surveillance Network (SSN) to track - typically objects larger than about 5-10 cm in diameter. There are expected to be even more debris objects at sizes too small to be seen and tracked by the SSN. Because of the altitude of the target satellite (865 x 845 km orbit), many of the debris are expected to have long orbital lifetimes and contribute to the orbital debris environment for decades to come. In the days and weeks following the ASAT test, NASA was able to use Lincoln Laboratory s Haystack radar on several occasions to observe portions of the ASAT debris cloud. Haystack has the capability of detecting objects down to less than one centimeter in diameter, and a large number of centimeter-sized particles corresponding to the ASAT cloud were clearly seen in the data. While Haystack cannot track these objects, the statistical sampling procedures NASA uses can give an accurate statistical picture of the characteristics of the debris from a breakup event. For years computer models based on data from ground hypervelocity collision tests (e.g., the SOCIT test) and orbital collision experiments (e.g., the P-78 and Delta-180 on-orbit collisions) have been used to predict the extent and characteristics of such hypervelocity collision debris clouds, but until now there have not been good ways to verify these models in the centimeter size regime. It is believed that unplanned collisions of objects in space similar to ASAT tests will drive the long-term future evolution of the debris environment in near-Earth space. Therefore, the Chinese ASAT test provides an excellent opportunity to test the models used to predict the future debris environment. For this study, Haystack detection events are compared to model predictions to test the model assumptions, including debris size distribution, velocity distribution, and assumptions about momentum transfer between the target and interceptor. In this paper we will present the results of these and other measurements on the size and extent of collisional breakup debris clouds.

  17. Recent Measurements of the Orbital Debris Environment at NASA Johnson Space Center

    NASA Technical Reports Server (NTRS)

    Stansbery, E. G.; Settecerri, T. J.; Africano, J. L.

    1999-01-01

    Space debris presents many challenges to current space operations. Although, the probability of collision between an operational spacecraft and a piece of space debris is quite small, the potential losses can be quite high. Prior to 1990, characterization of the orbital debris environment was divided into two categories. Objects larger than 10 cm are monitored by the United States Space Surveillance Network (SSN) and documented in the U.S. Space Command (USSPACECOM) catalog. Knowledge of debris smaller than 0.1 cm has come from the analyses of returned surfaces. The lack of information about the debris environment in the size range from 0.1 to 1 0 cm led to a joint NASA-DOD effort for orbital debris measurements using the Haystack radar and the unbuilt Haystack Auxiliary (HAX) radars. The data from these radars have been critical to the design of shielding for the International Space Station and have been extensively used in the creation of recent models describing the orbital debris environment. Recent debris campaigns have been conducted to verify and validate through comparative measurements, the results and conclusions drawn from the Haystack/HAX measurements. The Haystack/HAX measurements and results will be described as well as the results of the recent measurement campaigns.

  18. Finding and identifying the viral needle in the metagenomic haystack: trends and challenges

    PubMed Central

    Soueidan, Hayssam; Schmitt, Louise-Amélie; Candresse, Thierry; Nikolski, Macha

    2015-01-01

    Collectively, viruses have the greatest genetic diversity on Earth, occupy extremely varied niches and are likely able to infect all living organisms. Viral infections are an important issue for human health and cause considerable economic losses when agriculturally important crops or husbandry animals are infected. The advent of metagenomics has provided a precious tool to study viruses by sampling them in natural environments and identifying the genomic composition of a sample. However, reaching a clear recognition and taxonomic assignment of the identified viruses has been hampered by the computational difficulty of these problems. In this perspective paper we examine the trends in current research for the identification of viral sequences in a metagenomic sample, pinpoint the intrinsic computational difficulties for the identification of novel viral sequences within metagenomic samples, and suggest possible avenues to overcome them. PMID:25610431

  19. Haystack Ultrawideband Satellite Imaging Radar

    DTIC Science & Technology

    2014-09-01

    SEP 2014 2. REPORT TYPE 3. DATES COVERED 00-00-2014 to 00-00-2014 4. TITLE AND SUBTITLE Haystack Ultrawideband Satellite Imaging Radar 5a...www.ll.mit.edu September 2014 Since the launch of satellites into Earth orbits more than 50 years ago, space has become crowded. Commercial and military... satellites , both active and defunct, share the space environment with an assort- ment of space debris, such as remnants of damaged spacecraft and

  20. A Large-scale Traveling Ionospheric Disturbance during 2015 St. Patrick's Day Geomagnetic Storm

    NASA Astrophysics Data System (ADS)

    Liu, J.; Zhang, D.

    2016-12-01

    A large scale travelling ionospheric disturbance (LSTID) observed during the geomagnetic storm on St Patrick's Day, 2015 is studied with TEC data derived from the dense GPS network in East-Asia sector and, the TEC map from the Madrigal database of MIT Haystack Observatory; with the virtual height of the ionosphere at different frequencies from the China ionosonde network; and with HF Doppler shift observations in the mid-latitude sector of China. Results show that a LSTID spanning the 80°E-140°E longitude sector occurred propagating from higher to lower latitudes during the period between 9:40 and 11:30 UT. From the continuous two-dimensional TEC perturbation maps, the propagation features of this LSTID are studied. In the west side of China between 80°E-105°E, the LSTID propagates south by west; between 105°E-115°E, the LSTID propagates almost due south, and between 115°E-135°E, the LSTID propagates south by east. This feature of the propagation orientation appears to be related to the regional geomagnetic declination. The estimated velocity is from 531 m/s to 779 m/s, with the velocity in the east side of China larger than that estimated in the west side. Moreover, the velocity estimated from the movement of the wave front of LSTID valley is smaller than that estimated from the movement of the followed wave front of crest. In addition, the calculated wavelength of the LSTID is different between longitude sectors with the wavelength estimated to be about 2000 km near 115°E versus 1700 km near 95°E. Finally, using the TEC map data from the Madrigal database of the MIT Haystack Observatory, the characteristics of the ionospheric disturbances in European sector and in North American sector are also studied. The behavior of the ionospheric disturbance is very different between the East-Asian, European, and North American sectors. A longer lasting and continuous LSTID exists in the European sector (Greenwich Meridian line). The velocity of the LSTID along Greenwich Meridian line between the 70°N-60°N is smaller than that between 60 °N-30°N. In North American sector (80°W, 115°W), the equatorward LSTID is not obvious during the prevailing LSTID period studied here (between 9:40 and 11:30 UT).

  1. Formation of Ground Truth Databases and Related Studies and Regional Seismic Monitoring Research

    DTIC Science & Technology

    2006-06-01

    denoted by black circles) and those derived by the combined use of both IRIS and revised IDC for all 70 Chinese events. Arrows point towards the...IRIS data provide a large number of additional Lg amvals. Figure 45. Differences in epicentral location between the original REB (denoted by black ...Washington, USA Haystack Fork , Wyoming, USA Hells Canyon, Colorado, USA Hiawatha Road, Colorado, USA Hockley, Texas, USA Hailey, Idaho, USA Honiara

  2. Debris Flux Comparisons From The Goldstone Radar, Haystack Radar, and Hax Radar Prior, During, and After the Last Solar Maximum

    NASA Technical Reports Server (NTRS)

    Stokely, C. L.; Stansbery, E. G.; Goldstein, R. M.

    2006-01-01

    The continual monitoring of low Earth orbit (LEO) debris environment using highly sensitive radars is essential for an accurate characterization of these dynamic populations. Debris populations are continually evolving since there are new debris sources, previously unrecognized debris sources, and debris loss mechanisms that are dependent on the dynamic space environment. Such radar data are used to supplement, update, and validate existing orbital debris models. NASA has been utilizing radar observations of the debris environment for over a decade from three complementary radars: the NASA JPL Goldstone radar, the MIT Lincoln Laboratory (MIT/LL) Long Range Imaging Radar (known as the Haystack radar), and the MIT/LL Haystack Auxiliary radar (HAX). All of these systems are highly sensitive radars that operate in a fixed staring mode to statistically sample orbital debris in the LEO environment. Each of these radars is ideally suited to measure debris within a specific size region. The Goldstone radar generally observes objects with sizes from 2 mm to 1 cm. The Haystack radar generally measures from 5 mm to several meters. The HAX radar generally measures from 2 cm to several meters. These overlapping size regions allow a continuous measurement of cumulative debris flux versus diameter from 2 mm to several meters for a given altitude window. This is demonstrated for all three radars by comparing the debris flux versus diameter over 200 km altitude windows for 3 nonconsecutive years from 1998 through 2003. These years correspond to periods before, during, and after the peak of the last solar cycle. Comparing the year to year flux from Haystack for each of these altitude regions indicate statistically significant changes in subsets of the debris populations. Potential causes of these changes are discussed. These analysis results include error bars that represent statistical sampling errors, and are detailed in this paper.

  3. Haystack Observatory VLBI Correlator

    NASA Technical Reports Server (NTRS)

    Titus, Mike; Cappallo, Roger; Corey, Brian; Dudevoir, Kevin; Niell, Arthur; Whitney, Alan

    2013-01-01

    This report summarizes the activities of the Haystack Correlator during 2012. Highlights include finding a solution to the DiFX InfiniBand timeout problem and other DiFX software development, conducting a DBE comparison test following the First International VLBI Technology Workshop, conducting a Mark IV and DiFX correlator comparison, more broadband delay experiments, more u- VLBI Galactic Center observations, and conversion of RDV session processing to the Mark IV/HOPS path. Non-real-time e-VLBI transfers and engineering support of other correlators continued.

  4. Finding a needle in a haystack: toward a psychologically informed method for aviation security screening.

    PubMed

    Ormerod, Thomas C; Dando, Coral J

    2015-02-01

    Current aviation security systems identify behavioral indicators of deception to assess risks to flights, but they lack a strong psychological basis or empirical validation. We present a new method that tests the veracity of passenger accounts. In an in vivo double-blind randomized-control trial conducted in international airports, security agents detected 66% of deceptive passengers using the veracity test method compared with less than 5% using behavioral indicator recognition. As well as revealing advantages of veracity testing over behavioral indicator identification, the study provides the highest levels to date of deception detection in a realistic setting where the known base rate of deceptive individuals is low.

  5. Target identification for small bioactive molecules: finding the needle in the haystack.

    PubMed

    Ziegler, Slava; Pries, Verena; Hedberg, Christian; Waldmann, Herbert

    2013-03-04

    Identification and confirmation of bioactive small-molecule targets is a crucial, often decisive step both in academic and pharmaceutical research. Through the development and availability of several new experimental techniques, target identification is, in principle, feasible, and the number of successful examples steadily grows. However, a generic methodology that can successfully be applied in the majority of the cases has not yet been established. Herein we summarize current methods for target identification of small molecules, primarily for a chemistry audience but also the biological community, for example, the chemist or biologist attempting to identify the target of a given bioactive compound. We describe the most frequently employed experimental approaches for target identification and provide several representative examples illustrating the state-of-the-art. Among the techniques currently available, protein affinity isolation using suitable small-molecule probes (pulldown) and subsequent mass spectrometric analysis of the isolated proteins appears to be most powerful and most frequently applied. To provide guidance for rapid entry into the field and based on our own experience we propose a typical workflow for target identification, which centers on the application of chemical proteomics as the key step to generate hypotheses for potential target proteins. Copyright © 2013 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  6. NASA Technology Takes Center Stage

    NASA Technical Reports Server (NTRS)

    2004-01-01

    In today's fast-paced business world, there is often more information available to researchers than there is time to search through it. Data mining has become the answer to finding the proverbial "needle in a haystack," as companies must be able to quickly locate specific pieces of information from large collections of data. Perilog, a suite of data-mining tools, searches for hidden patterns in large databases to determine previously unrecognized relationships. By retrieving and organizing contextually relevant data from any sequence of terms - from genetic data to musical notes - the software can intelligently compile information about desired topics from databases.

  7. Needles in the EST Haystack: Large-Scale Identification and Analysis of Excretory-Secretory (ES) Proteins in Parasitic Nematodes Using Expressed Sequence Tags (ESTs)

    PubMed Central

    Nagaraj, Shivashankar H.; Gasser, Robin B.; Ranganathan, Shoba

    2008-01-01

    Background Parasitic nematodes of humans, other animals and plants continue to impose a significant public health and economic burden worldwide, due to the diseases they cause. Promising antiparasitic drug and vaccine candidates have been discovered from excreted or secreted (ES) proteins released from the parasite and exposed to the immune system of the host. Mining the entire expressed sequence tag (EST) data available from parasitic nematodes represents an approach to discover such ES targets. Methods and Findings In this study, we predicted, using EST2Secretome, a novel, high-throughput, computational workflow system, 4,710 ES proteins from 452,134 ESTs derived from 39 different species of nematodes, parasitic in animals (including humans) or plants. In total, 2,632, 786, and 1,292 ES proteins were predicted for animal-, human-, and plant-parasitic nematodes. Subsequently, we systematically analysed ES proteins using computational methods. Of these 4,710 proteins, 2,490 (52.8%) had orthologues in Caenorhabditis elegans, whereas 621 (13.8%) appeared to be novel, currently having no significant match to any molecule available in public databases. Of the C. elegans homologues, 267 had strong “loss-of-function” phenotypes by RNA interference (RNAi) in this nematode. We could functionally classify 1,948 (41.3%) sequences using the Gene Ontology (GO) terms, establish pathway associations for 573 (12.2%) sequences using Kyoto Encyclopaedia of Genes and Genomes (KEGG), and identify protein interaction partners for 1,774 (37.6%) molecules. We also mapped 758 (16.1%) proteins to protein domains including the nematode-specific protein family “transthyretin-like” and “chromadorea ALT,” considered as vaccine candidates against filariasis in humans. Conclusions We report the large-scale analysis of ES proteins inferred from EST data for a range of parasitic nematodes. This set of ES proteins provides an inventory of known and novel members of ES proteins as a foundation for studies focused on understanding the biology of parasitic nematodes and their interactions with their hosts, as well as for the development of novel drugs or vaccines for parasite intervention and control. PMID:18820748

  8. Assessing the usability of Shazam mobile app

    NASA Astrophysics Data System (ADS)

    Hussain, Azham; Mkpojiogu, Emmanuel O. C.; Almazini, Hassan; Almazini, Hussein

    2017-10-01

    Like searching for a "needle" in a "haystack", suppose that there is a large set of signals (finite sequences of numbers) fs1; s2; s3; g, and a special signal q that may or may not be in the collection. How can the signals be found in the collection that is similar, or identical to q, and how can this be done quickly? A solution to this question is the basis of the Shazam smart phone app, where a listener captures a short excerpt of a recorded song with the smart phone's microphone, and in a matter of moments the app reports the name of the song and the artist. Here, the "needle" is the excerpt, and the "haystack" is a vast corpus of popular music. The Shazam algorithm is powered by Fourier analysis. This paper presents a report on a usability evaluation of Amazon Shazam app. The researchers present the outcome based on a task-based evaluation that involved 15 users of different level of experience who performed 5 tasks using the Shazam mobile app. Post-test questionnaire was used to capture users' perceptions about the app. The results demonstrate that most of the participants were satisfied with services provided by the app.

  9. Titan's Stratospheric Condensibles at High Northern Latitudes During Northern Winter

    NASA Technical Reports Server (NTRS)

    Anderson, Carrie; Samuelson, R.; Achterberg, R.

    2012-01-01

    The Infrared Interferometer Spectrometer (IRIS) instrument on board Voyager 1 caught the first glimpse of an unidentified particulate feature in Titan's stratosphere that spectrally peaks at 221 per centimeter. Until recently, this feature that we have termed 'the haystack,' has been seen persistently at high northern latitudes with the Composite Infrared Spectrometer (CIRS) instrument onboard Cassini, The strength of the haystack emission feature diminishes rapidly with season, becoming drastically reduced at high northern latitudes, as Titan transitions from northern winter into spring, In contrast to IRIS whose shortest wavenumber was 200 per centimeter, CIRS extends down to 10 per centimeter, thus revealing an entirely unexplored spectral region in which nitrile ices have numerous broad lattice vibration features, Unlike the haystack, which is only found at high northern latitudes during northern winter/early northern spring, this geometrically thin nitrile cloud pervades Titan's lower stratosphere, spectrally peaking at 160 per centimeter, and is almost global in extent spanning latitudes 85 N to 600 S, The inference of nitrile ices are consistent with the highly restricted altitude ranges over which these features are observed, and appear to be dominated by a mixture of HCN and HC3N, The narrow range in altitude over which the nitrile ices extend is unlike the haystack, whose vertical distribution is significantly broader, spanning roughly 70 kilometers in altitude in Titan's lower stratosphere, The nitrile clouds that CIRS observes are located in a dynamically stable region of Titan's atmosphere, whereas CH4 clouds, which ordinarily form in the troposphere, form in a more dynamically unstable region, where convective cloud systems tend to occur. In the unusual situation where Titan's tropopause cools significantly from the HASI 70.5K temperature minimum, CH4 should condense in Titan's lower stratosphere, just like the aforementioned nitrile clouds, although in significantly larger abundances. We will present the spectral and vertical distribution of Titan's stratospheric particulates during northern winter on Titan. The drastically changing abundance of the haystack over a small latitude range will be highlighted, specifically comparing the IRIS and CIRS epochs, Finally, we will discuss the situation in which CH4 condenses in Titan's lower stratosphere, forming an unexpected quasi steady-state stratospheric Ice cloud.

  10. Information Content in Radio Waves: Student Investigations in Radio Science

    NASA Astrophysics Data System (ADS)

    Jacobs, K.; Scaduto, T.

    2013-12-01

    We describe an inquiry-based instructional unit on information content in radio waves, created in the summer of 2013 as part of a MIT Haystack Observatory (Westford, MA) NSF Research Experiences for Teachers (RET) program. This topic is current and highly relevant, addressing science and technical aspects from radio astronomy, geodesy, and atmospheric research areas as well as Next Generation Science Standards (NGSS). Projects and activities range from simple classroom demonstrations and group investigations, to long term research projects incorporating data acquisition from both student-built instrumentation as well as online databases. Each of the core lessons is applied to one of the primary research centers at Haystack through an inquiry project that builds on previously developed units through the MIT Haystack RET program. In radio astronomy, students investigate the application of a simple and inexpensive software defined radio chip (RTL-SDR) for use in systems implementing a small and very small radio telescope (SRT and VSRT). Both of these systems allow students to explore fundamental principles of radio waves and interferometry as applied to radio astronomy. In ionospheric research, students track solar storms from the initial coronal mass ejection (using Solar Dynamics Observatory images) to the resulting variability in total electron density concentrations using data from the community standard Madrigal distributed database system maintained by MIT Haystack. Finally, students get to explore very long-baseline interferometry as it is used in geodetic studies by measuring crustal plate displacements over time. Alignment to NextGen standards is provided for each lesson and activity with emphasis on HS-PS4 'Waves and Their Applications in Technologies for Information Transfer'.

  11. Post-Correlation Processing for the VLBI2010 Proof-of-Concept System

    NASA Technical Reports Server (NTRS)

    Beaudoin, Christopher; Niell, Arthur

    2010-01-01

    For the past three years, the MIT Haystack Observatory and the broadband team have been developing a proof-of-concept broadband geodetic VLBI microwave (2-12 GHz) receiver. Also on-going at Haystack is the development of post-correlation processing needed to extract the geodetic observables. Using this processing, the first fully-phase-calibrated geodetic fringes have been produced from observations conducted with the proof-of-concept system. The results we present show that the phase-calibrated phase residuals from four 512 MHz bands spanning 2 GHz have an RMS phase variation of 8deg which corresponds to a delay uncertainty of 12 ps.

  12. On the Determination of Poisson Statistics for Haystack Radar Observations of Orbital Debris

    NASA Technical Reports Server (NTRS)

    Stokely, Christopher L.; Benbrook, James R.; Horstman, Matt

    2007-01-01

    A convenient and powerful method is used to determine if radar detections of orbital debris are observed according to Poisson statistics. This is done by analyzing the time interval between detection events. For Poisson statistics, the probability distribution of the time interval between events is shown to be an exponential distribution. This distribution is a special case of the Erlang distribution that is used in estimating traffic loads on telecommunication networks. Poisson statistics form the basis of many orbital debris models but the statistical basis of these models has not been clearly demonstrated empirically until now. Interestingly, during the fiscal year 2003 observations with the Haystack radar in a fixed staring mode, there are no statistically significant deviations observed from that expected with Poisson statistics, either independent or dependent of altitude or inclination. One would potentially expect some significant clustering of events in time as a result of satellite breakups, but the presence of Poisson statistics indicates that such debris disperse rapidly with respect to Haystack's very narrow radar beam. An exception to Poisson statistics is observed in the months following the intentional breakup of the Fengyun satellite in January 2007.

  13. NASA Orbital Debris Large-Object Baseline Population in ORDEM 3.0

    NASA Technical Reports Server (NTRS)

    Krisco, Paula H.; Vavrin, A. B.; Anz-Meador, P. D.

    2013-01-01

    The NASA Orbital Debris Program Office (ODPO) has created and validated high fidelity populations of the debris environment for the latest Orbital Debris Engineering Model (ORDEM 3.0). Though the model includes fluxes of objects 10 um and larger, this paper considers particle fluxes for 1 cm and larger debris objects from low Earth orbit (LEO) through Geosynchronous Transfer Orbit (GTO). These are validated by several reliable radar observations through the Space Surveillance Network (SSN), Haystack, and HAX radars. ORDEM 3.0 populations were designed for the purpose of assisting, debris researchers and sensor developers in planning and testing. This environment includes a background derived from the LEO-to-GEO ENvironment Debris evolutionary model (LEGEND) with a Bayesian rescaling as well as specific events such as the FY-1C anti-satellite test, the Iridium 33/Cosmos 2251 accidental collision, and the Soviet/Russian Radar Ocean Reconnaissance Satellite (RORSAT) sodium-potassium droplet releases. The environment described in this paper is the most realistic orbital debris population larger than 1 cm, to date. We describe derivations of the background population and added specific populations. We present sample validation charts of our 1 cm and larger LEO population against Space Surveillance Network (SSN), Haystack, and HAX radar measurements.

  14. Spectral line and continuum studies using Haystack antenna

    NASA Technical Reports Server (NTRS)

    1973-01-01

    During the last half of 1972, the Haystack antenna was utilized 88% of the time. Of this useful time, 81% was devoted to radio astronomy investigations, 8% was spent on radar-related research and 11% was scheduled for maintenance and system improvements. Thirteen programs were completed of which 10 were spectral-line studies involving primarily recombination lines and H2O vapor investigations. The others involved 2 cm and 1.3 cm continuum observations. Fifteen new programs were accepted and the currently active radio observing programs totalled 24 as of 31 December 1973. The last radar measurements in the lunar topography program have now been completed. Radar activity, including measurements on Mercury, Venus and synchronous satellites has continued.

  15. Adults and Adolescents: Ambivalence and Ambiguity

    ERIC Educational Resources Information Center

    Staneck, Lou Willet

    1974-01-01

    Examines the ways in which adolescent books portrey adult characters. Some of the specific books covered are: Lisa, Bright and Dark; Island of the Blue Dolphins; Under the Haystack; and The Outsiders. (JB)

  16. Seasonal variability of mesospheric water vapor

    NASA Technical Reports Server (NTRS)

    Schwartz, P. R.; Bevilacqua, R. M.; Wilson, W. J.; Ricketts, W. B.; Howard, R. J.

    1985-01-01

    Ground-based spectral line measurements of the 22.2 GHz atmospheric water vapor line in emission were made at the JPL in order to obtain data in a dry climate, and to confirm similar measurements made at the Haystack Observatory. The results obtained from March 1984 to July 1984 and from December 1984 to May 1985, were based on data recorded by a HP9816 microcomputer. The instrument spectrometer was a 64 channel, 62.5 kHz resolution filter bank. Data indicates the existence of a seasonal variation in the abundance of water vapor in the upper mesosphere, with mixing ratios higher in summer than in spring. This is consistent with recent theoretical and observational results. In the area of semiannual oscillation, Haystack data are more consistent than those of JPL, indicating an annual cycle with abundances at maximum in summer and minimum in winter.

  17. A weak-scattering model for turbine-tone haystacking

    NASA Astrophysics Data System (ADS)

    McAlpine, A.; Powles, C. J.; Tester, B. J.

    2013-08-01

    Noise and emissions are critical technical issues in the development of aircraft engines. This necessitates the development of accurate models to predict the noise radiated from aero-engines. Turbine tones radiated from the exhaust nozzle of a turbofan engine propagate through turbulent jet shear layers which causes scattering of sound. In the far-field, measurements of the tones may exhibit spectral broadening, where owing to scattering, the tones are no longer narrow band peaks in the spectrum. This effect is known colloquially as 'haystacking'. In this article a comprehensive analytical model to predict spectral broadening for a tone radiated through a circular jet, for an observer in the far field, is presented. This model extends previous work by the authors which considered the prediction of spectral broadening at far-field observer locations outside the cone of silence. The modelling uses high-frequency asymptotic methods and a weak-scattering assumption. A realistic shear layer velocity profile and turbulence characteristics are included in the model. The mathematical formulation which details the spectral broadening, or haystacking, of a single-frequency, single azimuthal order turbine tone is outlined. In order to validate the model, predictions are compared with experimental results, albeit only at polar angle equal to 90°. A range of source frequencies from 4 to 20kHz, and jet velocities from 20 to 60ms-1, are examined for validation purposes. The model correctly predicts how the spectral broadening is affected when the source frequency and jet velocity are varied.

  18. VLBI geodesy - 2 parts-per-billion precision in length determinations for transcontinental baselines

    NASA Technical Reports Server (NTRS)

    Davis, J. L.; Herring, T. A.; Shapiro, I. I.

    1988-01-01

    VLBI was to make twenty-two independent measurements, between September 1984 and December 1986, of the length of the 3900-km baseline between the Mojave site in California and the Haystack/Westford site in Massachusetts. These experiments differ from the typical geodetic VLBI experiments in that a large fraction of observations is obtained at elevation angles between 4 and 10 deg. Data from these low elevation angles allow the vertical coordinate of site position, and hence the baseline length, to be estimated with greater precision. For the sixteen experiments processed thus far, the weighted root-mean-square scatter of the estimates of the baseline length is 8 mm.

  19. 1. AERIAL VIEW OF BOTH ACTIVE AND INACTIVE FLUMES, TAKEN ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    1. AERIAL VIEW OF BOTH ACTIVE AND INACTIVE FLUMES, TAKEN FROM EAST, NOTE STOCK SHELTERS IN BACKGROUND AND HAYSTACKS AND STORAGE IN FOREGROUND - Grant-Kohrs Ranch, Flumes, Highway 10, Deer Lodge, Powell County, MT

  20. Solid-State Chemistry as a Formation Mechanism for C 4N 2 Ice and Possibly the Haystack (220 cm -1 ice emission feature) in Titan's Stratosphere as Observed by Cassini CIRS

    NASA Astrophysics Data System (ADS)

    Anderson, Carrie; Samuelson, Robert E.; McLain, Jason L.; Nna Mvondo, Delphine; Romani, Paul; Flasar, F. Michael

    2016-10-01

    A profusion of organic ices containing hydrocarbons, nitriles, and combinations of their mixtures comprise Titan's complex stratospheric cloud systems, and are typically formed via vapor condensation. These ice particles are then distributed throughout the mid-to-lower stratosphere, with an increased abundance near the winter poles (see Anderson et al., 2016). The cold temperatures and the associated strong circumpolar winds that isolate polar air act in much the same way as on Earth, giving rise to compositional anomalies and stratospheric clouds that provide heterogeneous chemistry sites.Titan's C4N2 ice emission feature at 478 cm-1 and "the Haystack," a strong unidentified stratospheric ice emission feature centered at 220 cm-1, share a common characteristic. Even though both are distinctive ice emission features evident in Cassini Composite InfraRed (CIRS) far-IR spectra, no associated vapor emission features can be found in Titan's atmosphere. Without a vapor phase, solid-state chemistry provides an alternate mechanism beside vapor condensation for producing these observed stratospheric ices.Anderson et al., (2016) postulated that C4N2 ice formed in Titan's stratosphere via the solid-state photochemical reaction HCN + HC3N → C4N2 + H2 can occur within extant HCN-HC3N composite ice particles. Such a reaction, and potentially similar reactions that may produce the Haystack ice, are specific examples of solid-state chemistry in solar system atmospheres. This is in addition to the reaction HCl + ClONO2 → HNO3 + Cl2, which is known to produce HNO3 coatings on terrestrial water ice particles, a byproduct of the catalytic chlorine chemistry that produces ozone holes in Earth's polar stratosphere (see for example, Molina et al., 1987 Soloman, 1999).A combination of radiative transfer modeling of CIRS far-IR spectra, coupled with optical constants derived from thin film transmittance spectra of organic ice mixtures obtained in our Spectroscopy for Planetary ICes Environments (SPICE) laboratory, will be used to: 1) derive the vertical column abundance of C4N2 ice in Titan's early spring polar stratosphere, and 2) narrow the range of possible chemical compositions for the material comprising the Haystack.

  1. Sound production due to large-scale coherent structures. [and identification of noise mechanisms in turbulent shear flow

    NASA Technical Reports Server (NTRS)

    Gatski, T. B.

    1979-01-01

    The sound due to the large-scale (wavelike) structure in an infinite free turbulent shear flow is examined. Specifically, a computational study of a plane shear layer is presented, which accounts, by way of triple decomposition of the flow field variables, for three distinct component scales of motion (mean, wave, turbulent), and from which the sound - due to the large-scale wavelike structure - in the acoustic field can be isolated by a simple phase average. The computational approach has allowed for the identification of a specific noise production mechanism, viz the wave-induced stress, and has indicated the effect of coherent structure amplitude and growth and decay characteristics on noise levels produced in the acoustic far field.

  2. pGlyco 2.0 enables precision N-glycoproteomics with comprehensive quality control and one-step mass spectrometry for intact glycopeptide identification.

    PubMed

    Liu, Ming-Qi; Zeng, Wen-Feng; Fang, Pan; Cao, Wei-Qian; Liu, Chao; Yan, Guo-Quan; Zhang, Yang; Peng, Chao; Wu, Jian-Qiang; Zhang, Xiao-Jin; Tu, Hui-Jun; Chi, Hao; Sun, Rui-Xiang; Cao, Yong; Dong, Meng-Qiu; Jiang, Bi-Yun; Huang, Jiang-Ming; Shen, Hua-Li; Wong, Catherine C L; He, Si-Min; Yang, Peng-Yuan

    2017-09-05

    The precise and large-scale identification of intact glycopeptides is a critical step in glycoproteomics. Owing to the complexity of glycosylation, the current overall throughput, data quality and accessibility of intact glycopeptide identification lack behind those in routine proteomic analyses. Here, we propose a workflow for the precise high-throughput identification of intact N-glycopeptides at the proteome scale using stepped-energy fragmentation and a dedicated search engine. pGlyco 2.0 conducts comprehensive quality control including false discovery rate evaluation at all three levels of matches to glycans, peptides and glycopeptides, improving the current level of accuracy of intact glycopeptide identification. The N-glycoproteome of samples metabolically labeled with 15 N/ 13 C were analyzed quantitatively and utilized to validate the glycopeptide identification, which could be used as a novel benchmark pipeline to compare different search engines. Finally, we report a large-scale glycoproteome dataset consisting of 10,009 distinct site-specific N-glycans on 1988 glycosylation sites from 955 glycoproteins in five mouse tissues.Protein glycosylation is a heterogeneous post-translational modification that generates greater proteomic diversity that is difficult to analyze. Here the authors describe pGlyco 2.0, a workflow for the precise one step identification of intact N-glycopeptides at the proteome scale.

  3. Making Choices in Functional Vision Evaluations: "Noodles, Needles, and Haystacks."

    ERIC Educational Resources Information Center

    Bishop, V. E.

    1988-01-01

    An approach to functional vision evaluations clarifies the types of data collection and suggests protocols for three broad categories of visually handicapped children: "normal" school-age students, "normal" preschoolers, and multiply handicapped pupils. Visually impaired infants are classified with multiply handicapped pupils…

  4. Late Mesozoic and possible early Tertiary accretion in western Washington State: the Helena-Haystack melange and the Darrington- Devils Mountain fault zone

    USGS Publications Warehouse

    Tabor, R.W.

    1994-01-01

    The Helena-Haystack melange (HH melange) and coincident Darrington-Devils Mountain fault zone (DDMFZ) in northwestern Washington separate two terranes, the northwest Cascade System (NWCS) and the western and eastern melange belts (WEMB). The two terranes of Paleozoic and Mesozoic rocks superficially resemble each other but record considerable differences in structural and metamorphic history. The HH melange is a serpentinite-matrix melange containing blocks of adjacent terranes but also exotic blocks. The HH melange must have formed between early Cretaceous and late middle Eocene time, because it contains tectonic clasts of early Cretaceous Shuksan Greenschist and is overlain by late middle Eocene sedimentary and volcanic rocks. The possible continuation of the DDMFZ to the northwest as the San Juan and the West Coast faults on Vancouver Island suggests that the structure has had a major role in the emplacement of all the westernmost terranes in the Pacific Northwest. -from Author

  5. Test-state approach to the quantum search problem

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sehrawat, Arun; Nguyen, Le Huy; Graduate School for Integrative Sciences and Engineering, National University of Singapore, Singapore 117597

    2011-05-15

    The search for 'a quantum needle in a quantum haystack' is a metaphor for the problem of finding out which one of a permissible set of unitary mappings - the oracles - is implemented by a given black box. Grover's algorithm solves this problem with quadratic speedup as compared with the analogous search for 'a classical needle in a classical haystack'. Since the outcome of Grover's algorithm is probabilistic - it gives the correct answer with high probability, not with certainty - the answer requires verification. For this purpose we introduce specific test states, one for each oracle. These testmore » states can also be used to realize 'a classical search for the quantum needle' which is deterministic - it always gives a definite answer after a finite number of steps - and 3.41 times as fast as the purely classical search. Since the test-state search and Grover's algorithm look for the same quantum needle, the average number of oracle queries of the test-state search is the classical benchmark for Grover's algorithm.« less

  6. Hierarchical Learning of Tree Classifiers for Large-Scale Plant Species Identification.

    PubMed

    Fan, Jianping; Zhou, Ning; Peng, Jinye; Gao, Ling

    2015-11-01

    In this paper, a hierarchical multi-task structural learning algorithm is developed to support large-scale plant species identification, where a visual tree is constructed for organizing large numbers of plant species in a coarse-to-fine fashion and determining the inter-related learning tasks automatically. For a given parent node on the visual tree, it contains a set of sibling coarse-grained categories of plant species or sibling fine-grained plant species, and a multi-task structural learning algorithm is developed to train their inter-related classifiers jointly for enhancing their discrimination power. The inter-level relationship constraint, e.g., a plant image must first be assigned to a parent node (high-level non-leaf node) correctly if it can further be assigned to the most relevant child node (low-level non-leaf node or leaf node) on the visual tree, is formally defined and leveraged to learn more discriminative tree classifiers over the visual tree. Our experimental results have demonstrated the effectiveness of our hierarchical multi-task structural learning algorithm on training more discriminative tree classifiers for large-scale plant species identification.

  7. Oklahoma Fish Kill Study: Looking for a Toxic Needle in an Environmental Haystack

    EPA Science Inventory

    Since December 2011 the U.S. Environmental Protection Agency’s Office of Research and Development, National Exposure Research Laboratory-Environmental Sciences Division (EPA/ORD-NERL-ESD) has assisted EPA Region 6 and the State of Oklahoma Department of Environmental Quality (OKD...

  8. Competitive code-based fast palmprint identification using a set of cover trees

    NASA Astrophysics Data System (ADS)

    Yue, Feng; Zuo, Wangmeng; Zhang, David; Wang, Kuanquan

    2009-06-01

    A palmprint identification system recognizes a query palmprint image by searching for its nearest neighbor from among all the templates in a database. When applied on a large-scale identification system, it is often necessary to speed up the nearest-neighbor searching process. We use competitive code, which has very fast feature extraction and matching speed, for palmprint identification. To speed up the identification process, we extend the cover tree method and propose to use a set of cover trees to facilitate the fast and accurate nearest-neighbor searching. We can use the cover tree method because, as we show, the angular distance used in competitive code can be decomposed into a set of metrics. Using the Hong Kong PolyU palmprint database (version 2) and a large-scale palmprint database, our experimental results show that the proposed method searches for nearest neighbors faster than brute force searching.

  9. The needle in the hay stack: Keeping it lost

    USDA-ARS?s Scientific Manuscript database

    A recent advertisement on American commercial television depicts a well-dressed young man reaching into a haystack. In less than a second he removes his hand from the straw and briefly examines something held between his thumb and forefinger. With feigned surprise and a slight smirk, he quips, “Huh,...

  10. Worldwide time and frequency synchronization by planned VLBI networks

    NASA Technical Reports Server (NTRS)

    Coates, R. J.; Clark, T. A.

    1974-01-01

    Accurate baseline determinations and clock synchronization results obtained from the Quasar Patrol observations at X band with the Goldstone-Haystack baseline are presented. In addition, data from stations at Greenbank, West Virginia, and Onsala, Sweden were used. It was estimated that clock accuracy was on the order of 16 cm.

  11. Effort and Potential Efficiencies for Aquatic Non-native Species Early Detection

    EPA Science Inventory

    This manuscript is based on the early aquatic non-native species detection research in the Duluth-Superior harbor. The problem of early detection is essentially that of a "needle in a haystack" - to detect a newly arrived and presumably rare non-native species with a high probabi...

  12. Swarm Verification

    NASA Technical Reports Server (NTRS)

    Holzmann, Gerard J.; Joshi, Rajeev; Groce, Alex

    2008-01-01

    Reportedly, supercomputer designer Seymour Cray once said that he would sooner use two strong oxen to plow a field than a thousand chickens. Although this is undoubtedly wise when it comes to plowing a field, it is not so clear for other types of tasks. Model checking problems are of the proverbial "search the needle in a haystack" type. Such problems can often be parallelized easily. Alas, none of the usual divide and conquer methods can be used to parallelize the working of a model checker. Given that it has become easier than ever to gain access to large numbers of computers to perform even routine tasks it is becoming more and more attractive to find alternate ways to use these resources to speed up model checking tasks. This paper describes one such method, called swarm verification.

  13. Finding the Cold Needle in a Warm Haystack: Infrared Imaging Applied to Locating Cryo-cooled Crystals in Loops

    NASA Technical Reports Server (NTRS)

    Snell, Edward; vanderWoerd, Mark

    2003-01-01

    Thermally imaging the cryocooling processes of crystals has been demonstrated showing the progression of a cold wave through a crystal from the face closest to the origin of the coldstream ending at the point furthest away. During these studies large volume crystals were clearly distinguished from the loop holding them. Large volume crystals, used for neutron studies, were chosen deliberately to enhance the imaging. The different infrared transmission and reflectance properties of the crystal in comparison to the cryo-protectant are thought to be the parameter that produces the contrast making the crystal visible. As an application of the technology to locating crystals, more small crystals of lysozyme and a bFGF/dna complex were cryo-protected and imaged in large loops. The crystals were clearly distinguished from the vitrified solution. In the case of the bFGF/dna complex the illumination had to be carefully manipulated to enable the crystal to be seen in the visible spectrum. These preliminary results will be presented along with advantages and disadvantages of the technique and a discussion of how it might be applied.

  14. DNA barcoding at riverscape scales: Assessing biodiversity among fishes of the genus Cottus (Teleostei) in northern Rocky Mountain streams

    Treesearch

    Michael K. Young; Kevin S. McKelvey; Kristine L. Pilgrim; Michael K. Schwartz

    2013-01-01

    There is growing interest in broad-scale biodiversity assessments that can serve as benchmarks for identifying ecological change. Genetic tools have been used for such assessments for decades, but spatial sampling considerations have largely been ignored. Here, we demonstrate how intensive sampling efforts across a large geographical scale can influence identification...

  15. The University of Mississippi Studies in English (1990), New Series Volume VIII.

    ERIC Educational Resources Information Center

    Fisher, Benjamin Franklin, IV, Ed.

    This book, a new series in literary criticism and English studies, presents 24 essays. Essays in the book are: "Melville and the Question of Camoes" (George Monteiro); "Calvin Daniel Yost, Jr." (Benjamin Franklin Fisher IV); "'The Haystack in the Floods': An Uncharacteristic Preraphaelite Poem" (Veronica M. S.…

  16. The Effectiveness of Internet-Controlled Astronomical Research Instrumentation for Education

    ERIC Educational Resources Information Center

    Pratap, Preethi; Salah, Joseph

    2004-01-01

    Over the last decade, remote instruments have become widely used in astronomy. Educational applications are more recent. This paper describes a program to bring radio astronomy into the undergraduate classroom through the use of a remote research-grade radio telescope, the MIT Haystack Observatory 37 m telescope. We examine the effectiveness of…

  17. Identification and measurement of shrub type vegetation on large scale aerial photography

    NASA Technical Reports Server (NTRS)

    Driscoll, R. S.

    1970-01-01

    Important range-shrub species were identified at acceptable levels of accuracy on large-scale 70 mm color and color infrared aerial photographs. Identification of individual shrubs was significantly higher, however, on color infrared. Photoscales smaller than 1:2400 had limited value except for mature individuals of relatively tall species, and then only if crown margins did not overlap and sharp contrast was evident between the species and background. Larger scale photos were required for low-growing species in dense stands. The crown cover for individual species was estimated from the aerial photos either with a measuring magnifier or a projected-scale micrometer. These crown cover measurements provide techniques for earth-resource analyses when used in conjunction with space and high-altitude remotely procured photos.

  18. Finding a Needle in the Haystack: Identification of Functionally Important Minority Phases in an Operating Battery.

    PubMed

    Zhang, Kai; Ren, Fang; Wang, Xuelong; Hu, Enyuan; Xu, Yahong; Yang, Xiao-Qing; Li, Hong; Chen, Liquan; Pianetta, Piero; Mehta, Apurva; Yu, Xiqian; Liu, Yijin

    2017-12-13

    The in-depth understanding of the minority phases' roles in functional materials, e.g., batteries, is critical for optimizing the system performance and the operational efficiency. Although the visualization of battery electrode under operating conditions has been demonstrated, the development of advanced data-mining approaches is still needed in order to identify minority phases and to understand their functionalities. The present study uses nanoscale X-ray spectromicroscopy to study a functional LiCoO 2 /Li battery pouch cell. The data-mining approaches developed herein were used to search through over 10 million X-ray absorption spectra that cover more than 100 active cathode particles. Two particles with unanticipated chemical fingerprints were identified and further analyzed, providing direct evidence and valuable insight into the undesired side reactions involving the cation dissolution and precipitation as well as the local overlithiation-caused subparticle domain deactivation. The data-mining approach described in this work is widely applicable to many other structurally complex and chemically heterogeneous systems, in which the secondary/minority phases could critically affect the overall performance of the system, well beyond battery research.

  19. Finding a needle in the haystack: the costs and cost-effectiveness of syphilis diagnosis and treatment during pregnancy to prevent congenital syphilis in Kalomo District of Zambia.

    PubMed

    Larson, Bruce A; Lembela-Bwalya, Deophine; Bonawitz, Rachael; Hammond, Emily E; Thea, Donald M; Herlihy, Julie

    2014-01-01

    In March 2012, The Elizabeth Glaser Pediatric AIDS Foundation trained maternal and child health workers in Southern Province of Zambia to use a new rapid syphilis test (RST) during routine antenatal care. A recent study by Bonawitz et al. (2014) evaluated the impact of this roll out in Kalomo District. This paper estimates the costs and cost-effectiveness from the provider's perspective under the actual conditions observed during the first year of the RST roll out. Information on materials used and costs were extracted from program records. A decision-analytic model was used to evaluate the costs (2012 USD) and cost-effectiveness. Basic parameters needed for the model were based on the results from the evaluation study. During the evaluation study, 62% of patients received a RST, and 2.8% of patients tested were positive (and 10.4% of these were treated). Even with very high RST sensitivity and specificity (98%), true prevalence of active syphilis would be substantially less (estimated at <0.7%). For 1,000 new ANC patients, costs of screening and treatment were estimated at $2,136, and the cost per avoided disability-adjusted-life year lost (DALY) was estimated at $628. Costs change little if all positives are treated (because prevalence is low and treatment costs are small), but the cost-per-DALY avoided falls to just $66. With full adherence to guidelines, costs increase to $3,174 per 1,000 patients and the cost-per-DALY avoided falls to $60. Screening for syphilis is only useful for reducing adverse birth outcomes if patients testing positive are actually treated. Even with very low prevalence of syphilis (a needle in the haystack), cost effectiveness improves dramatically if those found positive are treated; additional treatment costs little but DALYs avoided are substantial. Without treatment, the needle is essentially found and thrown back into the haystack.

  20. Large-scale database searching using tandem mass spectra: looking up the answer in the back of the book.

    PubMed

    Sadygov, Rovshan G; Cociorva, Daniel; Yates, John R

    2004-12-01

    Database searching is an essential element of large-scale proteomics. Because these methods are widely used, it is important to understand the rationale of the algorithms. Most algorithms are based on concepts first developed in SEQUEST and PeptideSearch. Four basic approaches are used to determine a match between a spectrum and sequence: descriptive, interpretative, stochastic and probability-based matching. We review the basic concepts used by most search algorithms, the computational modeling of peptide identification and current challenges and limitations of this approach for protein identification.

  1. Proceedings of the Annual Meeting (14th) Aquatic Plant Control Research Planning and Operations Review, Held at Lake Eufaula, Oklahoma on 26-29 November 1979.

    DTIC Science & Technology

    1980-10-01

    Development; Problem Identification and Assessment for Aquatic Plant Management; Natural Succession of Aquatic Plants; Large-Scale Operations Management Test...of Insects and Pathogens for Control of Waterhyacinth in Louisiana; Large-Scale Operations Management Test to Evaluate Prevention Methodology for...Control of Eurasian Watermilfoil in Washington; Large-Scale Operations Management Test Using the White Amur at Lake Conway, Florida; and Aquatic Plant Control Activities in the Panama Canal Zone.

  2. An Indexing Needle in an Intelligence Haystack: Methodological Approaches in Exploring the Documentation of British Military Intelligence

    ERIC Educational Resources Information Center

    Brunt, Rodney M.

    2007-01-01

    Experiences in researching the documentation of the intelligence (codename Ultra) produced by breaking Enigma at Government Code & Cypher School, Bletchley Park, 1939-45, are described. The major problems are identified and shown to lie in the obscurity of the associated processes, disguised as they were within general bureaucratic…

  3. Finding a (pine) needle in a haystack: chloroplast genome sequence divergence in rare and widespread pines

    Treesearch

    J.B. Whittall; J. Syring; M. Parks; J. Buenrostro; C. Dick; A. Liston; R. Cronn

    2010-01-01

    Critical to conservation efforts and other investigations at low taxonomic levels, DNA sequence data offer important insights into the distinctiveness, biogeographic partitioning, and evolutionary histories of species. The resolving power of DNA sequences is often limited by insufficient variability at the intraspecific level. This is particularly true of studies...

  4. Needles in the Haystack: Finding Content Worth Preparing for Workplace Learning with the KEP Model

    ERIC Educational Resources Information Center

    Thalmann, Stefan; Maier, Ronald

    2017-01-01

    Knowledge transfer between employees is a primary concern in organizations. Employees create or acquire content that partially represents knowledge. These knowledge elements are specific to the context in and for which they are created and rarely address the learning needs of other employees in other work situations. Organizations therefore need…

  5. Finding needles in haystacks: linking scientific names, reference specimens and molecular data for Fungi

    Treesearch

    C.L. Schoch; B. Robbertse; V. Robert; R.G. Haight; K. Kovacs; B. Leung; W. Meyer; R.H. Nilsson; K. Hughes; A.N. Miller; P.M. Kirk; K. Abarenkov; M.C. Aime; H.A. Ariyawansa; M. Bidartondo; T. Boekhout; B. Buyck; Q. Cai; J. Chen; A. Crespo; P.W. Crous; U. Damm; Z.W. De Beer; B.T.M. Dentinger; P.K. Divakar; M. Duenas; N. Feau; K. Fliegerova; M.A. Garcia; Z.-W. Ge; G.W. Griffith; J.Z. Groenewald; M. Groenewald; M. Grube; M. Gryzenhout; C. Gueidan; L. Guo; S. Hambleton; R. Hamelin; K. Hansen; V. Hofstetter; S.-B. Hong; J. Houbraken; K.D. Hyde; P. Inderbitzin; P.R. Johnston; S.C. Karunarathna; U. Koljalg; G.M. Kovacs; E. Kraichak; K. Krizsan; C.P. Kurtzman; K.-H. Larsson; S. Leavitt; P.M. Letcher; K. Liimatainen; J.-K. Liu; D.J. Lodge; J. Jennifer Luangsa-ard; H.T. Lumbsch; S.S.N. Maharachchikumbura; D. Manamgoda; M.P. Martin; A.M. Minnis; J.-M. Moncalvo; G. Mule; K.K. Nakasone; T. Niskanen; I. Olariaga; T. Papp; T. Petkovits; R. Pino-Bodas; M.J. Powell; H.A. Raja; D. Redecker; J.M. Sarmiento-Ramirez; K.A. Seifert; B. Shrestha; S. Stenroos; B. Stielow; S.-O. Suh; K. Tanaka; L. Tedersoo; M.T. Telleria; D. Udayanga; W.A. Untereiner; J. Dieguez Uribeondo; K.V. Subbarao; C. Vagvolgyi; C. Visagie; K. Voigt; D.M. Walker; B.S. Weir; M. Weiss; N.N. Wijayawardene; M.J. Wingfield; J.P. Xu; Z.L. Yang; N. Zhang; W.-Y. Zhuang; S. Federhen

    2014-01-01

    DNA phylogenetic comparisons have shown that morphology-based species recognition often underestimates fungal diversity. Therefore, the need for accurate DNA sequence data, tied to both correct taxonomic names and clearly annotated specimen data, has never been greater. Furthermore, the growing number of molecular ecology and microbiome projects using high-throughput...

  6. Looking for the Information Needle in the Internet Haystack.

    ERIC Educational Resources Information Center

    Clausen, Helge

    This paper presents the results of an empirical survey conducted in the winter of 1995-96 among 413 Danish Internet users (business people, higher education/research and students). Data from a recent Danish Mediacenter Internet survey (n=2981) are also included. Findings indicate that women tend to use the Internet to solve well-defined problems,…

  7. Needle in a Haystack? Seeking Causal Evidence about Using STEM Experts to Improve Student Outcomes

    ERIC Educational Resources Information Center

    Gamse, Beth C.; Martinez, Alina; Bozzi, Laurie

    2015-01-01

    In today's technological and global society, STEM disciplines are viewed as fundamental to the nation's economic growth and prosperity. Employment opportunities in STEM fields have increased at a faster rate than in non-STEM fields (Government Accounting Office [GAO], 2006). Additionally, many professions--once perceived not to require STEM…

  8. Task Scheduling in Desktop Grids: Open Problems

    NASA Astrophysics Data System (ADS)

    Chernov, Ilya; Nikitina, Natalia; Ivashko, Evgeny

    2017-12-01

    We survey the areas of Desktop Grid task scheduling that seem to be insufficiently studied so far and are promising for efficiency, reliability, and quality of Desktop Grid computing. These topics include optimal task grouping, "needle in a haystack" paradigm, game-theoretical scheduling, domain-imposed approaches, special optimization of the final stage of the batch computation, and Enterprise Desktop Grids.

  9. Finding a Needle in a Climate Haystack

    NASA Astrophysics Data System (ADS)

    Verosub, K. L.; Medrano, R.; Valentine, M.

    2014-12-01

    We are studying the regional impact of volcanic eruptions that might have caused global cooling using high-quality annual-resolution proxy records of natural phenomena, such as tree-ring widths, and cultural events, such as the dates of the beginning of grape and rye harvests. To do this we need to determine if the year following an eruption was significantly colder and wetter than preceding or subsequent years as measured by any given proxy and if that year is consistently cold and wet across different proxies. The problem is complicated by the fact that normal inter-annual variations in any given proxy can be quite large and can obscure any volcanological impact and by the fact that inter-annual variations for different proxies will have different means and standard deviations. We address the first problem by assuming that on a regional scale, the inter-annual variations of different proxies are at best only weakly correlated and that, in the absence of a volcanological signal, these variations will average out on a regional scale. We address the second problem by renormalizing each record so that it has the same mean and standard deviation over a given time interval. We then sum the re-normalized records on a year-by-year basis and look for years with significantly higher total scores. The method can also be used to assess the statistical significance of an anomalous value. Our initial analysis of records primarily from the Northern Hemisphere shows that the years 1601 and 1816 were significantly colder and wetter than any others in the past 500 years. These years followed the eruptions of Huayanaputina in Chile and Tambora in Indonesia, respectively, by one year. The years 1698 and 1837 also show up as being climatologically severe although they have not (yet) been associated with specific volcanic eruptions.

  10. Incorporation of DNA barcoding into a large-scale biomonitoring program: opportunities and pitfalls

    EPA Science Inventory

    Taxonomic identification of benthic macroinvertebrates is critical to protocols used to assess the biological integrity of aquatic ecosystems. The time, expense, and inherent error rate of species-level morphological identifications has necessitated use of genus- or family-level ...

  11. Target-decoy Based False Discovery Rate Estimation for Large-scale Metabolite Identification.

    PubMed

    Wang, Xusheng; Jones, Drew R; Shaw, Timothy I; Cho, Ji-Hoon; Wang, Yuanyuan; Tan, Haiyan; Xie, Boer; Zhou, Suiping; Li, Yuxin; Peng, Junmin

    2018-05-23

    Metabolite identification is a crucial step in mass spectrometry (MS)-based metabolomics. However, it is still challenging to assess the confidence of assigned metabolites. In this study, we report a novel method for estimating false discovery rate (FDR) of metabolite assignment with a target-decoy strategy, in which the decoys are generated through violating the octet rule of chemistry by adding small odd numbers of hydrogen atoms. The target-decoy strategy was integrated into JUMPm, an automated metabolite identification pipeline for large-scale MS analysis, and was also evaluated with two other metabolomics tools, mzMatch and mzMine 2. The reliability of FDR calculation was examined by false datasets, which were simulated by altering MS1 or MS2 spectra. Finally, we used the JUMPm pipeline coupled with the target-decoy strategy to process unlabeled and stable-isotope labeled metabolomic datasets. The results demonstrate that the target-decoy strategy is a simple and effective method for evaluating the confidence of high-throughput metabolite identification.

  12. A Needle in a Haystack: A Search for Value for Money in Turn-Around Strategies for Schools in Difficult Circumstances

    ERIC Educational Resources Information Center

    Khupe, Constance; Balkwill, Kevin; Osman, Ruksana; Cameron, Ann

    2013-01-01

    Background: While current investments in school improvement occur in the context of the worldwide economic downturn, in the South African context, there is in addition widespread disparity in education provision and attainment related to pre-democracy race-based patterns. Despite the education sector receiving the largest national annual budget…

  13. Of Needles and Haystacks: Building an Accurate Statewide Dropout Early Warning System in Wisconsin

    ERIC Educational Resources Information Center

    Knowles, Jared E.

    2015-01-01

    The state of Wisconsin has one of the highest four year graduation rates in the nation, but deep disparities among student subgroups remain. To address this the state has created the Wisconsin Dropout Early Warning System (DEWS), a predictive model of student dropout risk for students in grades six through nine. The Wisconsin DEWS is in use…

  14. Seminoe 3, a tidally influenced lowstand wedge and its relationships with subjacent highstand and overlying transgressive deposits, Haystack Mountains Formation, Cretaceous Western Interior, Wyoming (USA)

    NASA Astrophysics Data System (ADS)

    Mellere, Donatella

    1996-06-01

    The Campanian Haystack Mountains Formation (Wyoming, USA) consists of a series of deltaic/estuarine and shoreface sandstone tongues stacked in basinward-stepping units separated by shale intervals. Regional analysis indicates that the tongues have a distinctive two-tiered architecture: (a) a lower sandbody with bioturbated and hummocky strata of wave-dominated shoreface origin; and (b) an erosively based upper unit with bioturbated to tabular cross-bedded sandstones. This study focuses on a single tongue, Seminoe 3, which crops out in a 40 km long continuous dip-oriented exposure. In its landward reaches (westwards) the tongue consists of a thin unit of offshore-transitional to lower shoreface deposits. In the central part of the exposure it becomes thicker and consists of a cross-bedded wedge which truncates the shoreface deposits and pinches out by downlap eastwards onto a thick, westward migrating, shoreface lithosome (spit), probably developed on the margin of a wave-dominated delta. Detailed correlation and mapping of bounding surfaces indicate that the shoreface and spit lithosomes were deposited in a forced regressive mode during falling relative sea-level. The overlying cross-bedded lithosome, is a lowstand wedge which prograded over an incised topography created during the early stage of sea-level lowstand. A large embayment is suggested to have developed during the highstand/forced regressive phase, with shoreline and spit deposits prograding from opposing sides. Significant erosion and truncation occurred during the lowstand phase, as well as a basinward shift of facies onto the shelf and the pre-existing spit deposits. In the early stage of transgression the embayment was drowned, and a tide-dominated, estuarine environment was established. The resulting cross-bedded wedge displays an aggrading to prograding stacking pattern of parasequences with the uppermost portion exhibiting a retrogradational pattern, as a consequence of the reworking operated by the transgression.

  15. Large-scale DNA Barcode Library Generation for Biomolecule Identification in High-throughput Screens.

    PubMed

    Lyons, Eli; Sheridan, Paul; Tremmel, Georg; Miyano, Satoru; Sugano, Sumio

    2017-10-24

    High-throughput screens allow for the identification of specific biomolecules with characteristics of interest. In barcoded screens, DNA barcodes are linked to target biomolecules in a manner allowing for the target molecules making up a library to be identified by sequencing the DNA barcodes using Next Generation Sequencing. To be useful in experimental settings, the DNA barcodes in a library must satisfy certain constraints related to GC content, homopolymer length, Hamming distance, and blacklisted subsequences. Here we report a novel framework to quickly generate large-scale libraries of DNA barcodes for use in high-throughput screens. We show that our framework dramatically reduces the computation time required to generate large-scale DNA barcode libraries, compared with a naїve approach to DNA barcode library generation. As a proof of concept, we demonstrate that our framework is able to generate a library consisting of one million DNA barcodes for use in a fragment antibody phage display screening experiment. We also report generating a general purpose one billion DNA barcode library, the largest such library yet reported in literature. Our results demonstrate the value of our novel large-scale DNA barcode library generation framework for use in high-throughput screening applications.

  16. Efficacy and well-being in rural north India: The role of social identification with a large-scale community identity

    PubMed Central

    Khan, Sammyh S; Hopkins, Nick; Tewari, Shruti; Srinivasan, Narayanan; Reicher, Stephen David; Ozakinci, Gozde

    2014-01-01

    Identifying with a group can contribute to a sense of well-being. The mechanisms involved are diverse: social identification with a group can impact individuals' beliefs about issues such as their connections with others, the availability of social support, the meaningfulness of existence, and the continuity of their identity. Yet, there seems to be a common theme to these mechanisms: identification with a group encourages the belief that one can cope with the stressors one faces (which is associated with better well-being). Our research investigated the relationship between identification, beliefs about coping, and well-being in a survey (N = 792) administered in rural North India. Using structural equation modelling, we found that social identification as a Hindu had positive and indirect associations with three measures of well-being through the belief that one can cope with everyday stressors. We also found residual associations between participants' social identification as a Hindu and two measures of well-being in which higher identification was associated with poorer well-being. We discuss these findings and their implication for understanding the relationship between social identification (especially with large-scale group memberships) and well-being. We also discuss the application of social psychological theory developed in the urban West to rural north India. © 2014 The Authors. European Journal of Social Psychology published by John Wiley & Sons, Ltd. PMID:26160989

  17. Statistical characterization of the Sub-Auroral Polarization Stream (SAPS)

    NASA Astrophysics Data System (ADS)

    Kunduri, B.; Baker, J. B.; Ruohoniemi, J. M.; Erickson, P. J.; Coster, A. J.; Oksavik, K.

    2017-12-01

    The Sub-Auroral Polarization Stream (SAPS) is a narrow region of westward directed plasma convection typically observed in the dusk-midnight sector equatorward of the main auroral oval. SAPS plays an important role in mid-latitude space weather dynamics and has a controlling influence on the evolution of large-scale plasma features, such as Storm Enhanced Density (SED) plumes. In this study, data from North American mid-latitude SuperDARN radars collected between January 2011 and December 2014 have been used to compile a database of SAPS events for statistical analysis. We examine the dependence of SAPS velocity magnitude and direction on geomagnetic activity and magnetic local time. The lowest speed limit and electric fields observed during SAPS are discussed and histograms of SAPS velocities for different Dst bins and MLAT-MLT locations are presented. We find significant differences in SAPS characteristics between periods of low and high geomagnetic activity, suggesting that SAPS are driven by different mechanisms during storm and non-storm conditions. To further explore this possibility, we have characterized the SAPS location and peak speed relative to the ionospheric trough specified by GPS Total Electron Content (TEC) data from the MIT Haystack Madrigal database. A particular emphasis is placed on identifying the extent to which the location, structure, and depth of the trough may play a controlling influence on SAPS speeds during storm and non-storm periods. The results are interpreted in terms of the current paradigm for active thermosphere-ionosphere feedback being an important component of SAPS physics.

  18. Detecting rare, abnormally large grains by x-ray diffraction

    DOE PAGES

    Boyce, Brad L.; Furnish, Timothy Allen; Padilla, H. A.; ...

    2015-07-16

    Bimodal grain structures are common in many alloys, arising from a number of different causes including incomplete recrystallization and abnormal grain growth. These bimodal grain structures have important technological implications, such as the well-known Goss texture which is now a cornerstone for electrical steels. Yet our ability to detect bimodal grain distributions is largely confined to brute force cross-sectional metallography. The present study presents a new method for rapid detection of unusually large grains embedded in a sea of much finer grains. Traditional X-ray diffraction-based grain size measurement techniques such as Scherrer, Williamson–Hall, or Warren–Averbach rely on peak breadth andmore » shape to extract information regarding the average crystallite size. However, these line broadening techniques are not well suited to identify a very small fraction of abnormally large grains. The present method utilizes statistically anomalous intensity spikes in the Bragg peak to identify regions where abnormally large grains are contributing to diffraction. This needle-in-a-haystack technique is demonstrated on a nanocrystalline Ni–Fe alloy which has undergone fatigue-induced abnormal grain growth. In this demonstration, the technique readily identifies a few large grains that occupy <0.00001 % of the interrogation volume. Finally, while the technique is demonstrated in the current study on nanocrystalline metal, it would likely apply to any bimodal polycrystal including ultrafine grained and fine microcrystalline materials with sufficiently distinct bimodal grain statistics.« less

  19. Foreign Language Analysis and Recognition (FLARE) Progress

    DTIC Science & Technology

    2015-02-01

    Copies may be obtained from the Defense Technical Information Center (DTIC) (http://www.dtic.mil). AFRL- RH -WP-TR-2015-0007 HAS BEEN REVIEWED AND IS... retrieval (IR). 15. SUBJECT TERMS Automatic speech recognition (ASR), information retrieval (IR). 16. SECURITY CLASSIFICATION OF: 17. LIMITATION OF...to the Haystack Multilingual Multimedia Information Extraction and Retrieval (MMIER) system that was initially developed under a prior work unit

  20. News and Views: A VISTA of the Orion Nebula; Grote Reber Award; Leverhulme Prize; GNSS workshop; Farming in space; Space messages

    NASA Astrophysics Data System (ADS)

    2010-04-01

    The recipient of the 2010 Grote Reber Award is Dr Alan Rogers, a Research Affiliate at the Massachusetts Institute of Technology Haystack Observatory. Matt King, a Reader in Polar Geodesy in the School of Civil Engineering and Geosciences at the University of Newcastle, was awarded a Philip Leverhulme Prize in 2009.

  1. Missions and Medicine at Amherst: Family Ties to Edward Hitchcock Jr., the Missionary Movement, and the American University of Beirut

    ERIC Educational Resources Information Center

    Dorman, John M.

    2011-01-01

    The Haystack Movement began at Williams College in 1805, occasioning the spread of American missions throughout the world. A half century later, two graduates of nearby Amherst College, Edward Hitchcock Jr. and Daniel Bliss, laid the foundations for college health services in this country and for mission work and education in the Middle East. The…

  2. Finding a Needle in the Haystack: Identification of Functionally Important Minority Phases in an Operating Battery

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhang, Kai; Ren, Fang; Wang, Xuelong

    The in-depth understanding of the minority phases’ roles in functional materials, e.g., batteries, is critical for optimizing the system performance and the operational efficiency. Although the visualization of battery electrode under operating conditions has been demonstrated, the development of advanced data-mining approaches is still needed in order to identify minority phases and to understand their functionalities. The present study uses nanoscale X-ray spectromicroscopy to study a functional LiCoO 2/Li battery pouch cell. The data-mining approaches developed herein were used to search through over 10 million X-ray absorption spectra that cover more than 100 active cathode particles. Two particles with unanticipatedmore » chemical fingerprints were identified and further analyzed, providing direct evidence and valuable insight into the undesired side reactions involving the cation dissolution and precipitation as well as the local overlithiation-caused subparticle domain deactivation. As a result, the data-mining approach described in this work is widely applicable to many other structurally complex and chemically heterogeneous systems, in which the secondary/minority phases could critically affect the overall performance of the system, well beyond battery research.« less

  3. MotifMark: Finding regulatory motifs in DNA sequences.

    PubMed

    Hassanzadeh, Hamid Reza; Kolhe, Pushkar; Isbell, Charles L; Wang, May D

    2017-07-01

    The interaction between proteins and DNA is a key driving force in a significant number of biological processes such as transcriptional regulation, repair, recombination, splicing, and DNA modification. The identification of DNA-binding sites and the specificity of target proteins in binding to these regions are two important steps in understanding the mechanisms of these biological activities. A number of high-throughput technologies have recently emerged that try to quantify the affinity between proteins and DNA motifs. Despite their success, these technologies have their own limitations and fall short in precise characterization of motifs, and as a result, require further downstream analysis to extract useful and interpretable information from a haystack of noisy and inaccurate data. Here we propose MotifMark, a new algorithm based on graph theory and machine learning, that can find binding sites on candidate probes and rank their specificity in regard to the underlying transcription factor. We developed a pipeline to analyze experimental data derived from compact universal protein binding microarrays and benchmarked it against two of the most accurate motif search methods. Our results indicate that MotifMark can be a viable alternative technique for prediction of motif from protein binding microarrays and possibly other related high-throughput techniques.

  4. Finding a Needle in the Haystack: Identification of Functionally Important Minority Phases in an Operating Battery

    DOE PAGES

    Zhang, Kai; Ren, Fang; Wang, Xuelong; ...

    2017-11-08

    The in-depth understanding of the minority phases’ roles in functional materials, e.g., batteries, is critical for optimizing the system performance and the operational efficiency. Although the visualization of battery electrode under operating conditions has been demonstrated, the development of advanced data-mining approaches is still needed in order to identify minority phases and to understand their functionalities. The present study uses nanoscale X-ray spectromicroscopy to study a functional LiCoO 2/Li battery pouch cell. The data-mining approaches developed herein were used to search through over 10 million X-ray absorption spectra that cover more than 100 active cathode particles. Two particles with unanticipatedmore » chemical fingerprints were identified and further analyzed, providing direct evidence and valuable insight into the undesired side reactions involving the cation dissolution and precipitation as well as the local overlithiation-caused subparticle domain deactivation. As a result, the data-mining approach described in this work is widely applicable to many other structurally complex and chemically heterogeneous systems, in which the secondary/minority phases could critically affect the overall performance of the system, well beyond battery research.« less

  5. Uncovering Special Nuclear Materials by Low-energy Nuclear Reaction Imaging.

    PubMed

    Rose, P B; Erickson, A S; Mayer, M; Nattress, J; Jovanovic, I

    2016-04-18

    Weapons-grade uranium and plutonium could be used as nuclear explosives with extreme destructive potential. The problem of their detection, especially in standard cargo containers during transit, has been described as "searching for a needle in a haystack" because of the inherently low rate of spontaneous emission of characteristic penetrating radiation and the ease of its shielding. Currently, the only practical approach for uncovering well-shielded special nuclear materials is by use of active interrogation using an external radiation source. However, the similarity of these materials to shielding and the required radiation doses that may exceed regulatory limits prevent this method from being widely used in practice. We introduce a low-dose active detection technique, referred to as low-energy nuclear reaction imaging, which exploits the physics of interactions of multi-MeV monoenergetic photons and neutrons to simultaneously measure the material's areal density and effective atomic number, while confirming the presence of fissionable materials by observing the beta-delayed neutron emission. For the first time, we demonstrate identification and imaging of uranium with this novel technique using a simple yet robust source, setting the stage for its wide adoption in security applications.

  6. Combinatorial Approach for Large-scale Identification of Linked Peptides from Tandem Mass Spectrometry Spectra*

    PubMed Central

    Wang, Jian; Anania, Veronica G.; Knott, Jeff; Rush, John; Lill, Jennie R.; Bourne, Philip E.; Bandeira, Nuno

    2014-01-01

    The combination of chemical cross-linking and mass spectrometry has recently been shown to constitute a powerful tool for studying protein–protein interactions and elucidating the structure of large protein complexes. However, computational methods for interpreting the complex MS/MS spectra from linked peptides are still in their infancy, making the high-throughput application of this approach largely impractical. Because of the lack of large annotated datasets, most current approaches do not capture the specific fragmentation patterns of linked peptides and therefore are not optimal for the identification of cross-linked peptides. Here we propose a generic approach to address this problem and demonstrate it using disulfide-bridged peptide libraries to (i) efficiently generate large mass spectral reference data for linked peptides at a low cost and (ii) automatically train an algorithm that can efficiently and accurately identify linked peptides from MS/MS spectra. We show that using this approach we were able to identify thousands of MS/MS spectra from disulfide-bridged peptides through comparison with proteome-scale sequence databases and significantly improve the sensitivity of cross-linked peptide identification. This allowed us to identify 60% more direct pairwise interactions between the protein subunits in the 20S proteasome complex than existing tools on cross-linking studies of the proteasome complexes. The basic framework of this approach and the MS/MS reference dataset generated should be valuable resources for the future development of new tools for the identification of linked peptides. PMID:24493012

  7. An accurate and computationally efficient algorithm for ground peak identification in large footprint waveform LiDAR data

    NASA Astrophysics Data System (ADS)

    Zhuang, Wei; Mountrakis, Giorgos

    2014-09-01

    Large footprint waveform LiDAR sensors have been widely used for numerous airborne studies. Ground peak identification in a large footprint waveform is a significant bottleneck in exploring full usage of the waveform datasets. In the current study, an accurate and computationally efficient algorithm was developed for ground peak identification, called Filtering and Clustering Algorithm (FICA). The method was evaluated on Land, Vegetation, and Ice Sensor (LVIS) waveform datasets acquired over Central NY. FICA incorporates a set of multi-scale second derivative filters and a k-means clustering algorithm in order to avoid detecting false ground peaks. FICA was tested in five different land cover types (deciduous trees, coniferous trees, shrub, grass and developed area) and showed more accurate results when compared to existing algorithms. More specifically, compared with Gaussian decomposition, the RMSE ground peak identification by FICA was 2.82 m (5.29 m for GD) in deciduous plots, 3.25 m (4.57 m for GD) in coniferous plots, 2.63 m (2.83 m for GD) in shrub plots, 0.82 m (0.93 m for GD) in grass plots, and 0.70 m (0.51 m for GD) in plots of developed areas. FICA performance was also relatively consistent under various slope and canopy coverage (CC) conditions. In addition, FICA showed better computational efficiency compared to existing methods. FICA's major computational and accuracy advantage is a result of the adopted multi-scale signal processing procedures that concentrate on local portions of the signal as opposed to the Gaussian decomposition that uses a curve-fitting strategy applied in the entire signal. The FICA algorithm is a good candidate for large-scale implementation on future space-borne waveform LiDAR sensors.

  8. Making Sense of Minority Student Identification in Special Education: School Context Matters

    ERIC Educational Resources Information Center

    Talbott, Elizabeth; Fleming, Jane; Karabatsos, George; Dobria, Lidia

    2011-01-01

    Since the inception of special education, researchers have identified higher proportions of minority students with disabilities than expected. Yet, relatively few studies have considered the contributions of the school context on a large scale to the identification of students with mental retardation (MR), emotional disturbance (ED), and learning…

  9. Searches for and identification of effects of extra spatial dimensions in dilepton and diphoton production at the Large Hadron Collider

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pankov, A. A., E-mail: pankov@ictp.it; Serenkova, I. A., E-mail: inna.serenkova@cern.ch; Tsytrinov, A. V., E-mail: tsytrin@gstu.by

    2015-06-15

    Prospects of discovering and identifying effects of extra spatial dimensions in dilepton and diphoton production at the Large Hadron Collider (LHC) are studied. Such effects may be revealed by the characteristic behavior of the invariant-mass distributions of dileptons and diphotons, and their identification can be performed on the basis of an analysis of their angular distributions. The discovery and identification reaches are estimated for the scale parameter M{sub S} of the Kaluza-Klein gravitational towers, which can be determined in experiments devoted to measuring the dilepton and diphoton channels at the LHC.

  10. Merced County Streams Project, Haystack Reservoir, California Intensive Cultural Resources Survey.

    DTIC Science & Technology

    1982-03-25

    evaluation of sampling strategies: simulated " 1excavations of a Kenyan pastoralist site. In Simulation studies in archaeology , edited by Ian Hodder, pp. 123...Publications ift American Archaeology and Ethnology 6(2):333-368. Berkeley. Barrett, Samuel A., and Edward W. Gifford _ -1933 Miwok material culture...University of Oregon Books, Eugene. !r Bennyhoff, James A. 1956 An appraisal of the archaeological resources of Yosemite National Park. University of

  11. Rethinking Intelligence to Integrate Counterterrorism into the Local Law Enforcement Mission

    DTIC Science & Technology

    2007-03-01

    a needle in the haystack problem. Also referred to as the wheat versus the chaff problem, valuable information must be separated from unimportant...information and processed before analysts can yield any useful intelligence.25 3. Processing and Exploitation To address the wheat -versus-chaff...93 Despite the perception that Chicago is an aging Rust Belt city, some experts report that it has the largest high technology and information

  12. Updating the NASA LEO Orbital Debris Environment Model with Recent Radar and Optical Observations and in Situ Measurements

    NASA Technical Reports Server (NTRS)

    Liou, J.-C.; Anz-Meador, P.; Matney, M. J.; Kessler, D. J.; Theall, J.; Johnson, N. L.

    2000-01-01

    The Low Earth Orbit (LEO, between 200 and 2000 km altitudes) debris environment has been constantly measured by NASA Johnson Space Center's Liquid Mirror Telescope (LMT) since 1996 (Africano et al. 1999, NASA JSC-28826) and by Haystack and Haystack Auxiliary radars at MIT Lincoln Laboratory since 1990 (Settecerri et al. 1999, NASA JSC-28744). Debris particles as small as 3 mm can be detected by the radars and as small as 3 cm can be measured by LMT. Objects about 10 cm in diameter and greater are tracked and catalogued by the US Space Surveillance Network. Much smaller (down to several micrometers) natural and debris particle populations can be estimated based on in situ measurements, such as Long Duration Exposure Facility, and based on analyses of returned surfaces, such as Hubble Space Telescope solar arrays, European Retrievable Carrier, and Space Shuttles. To increase our understanding of the current LEO debris environment, the Orbital Debris Program Office at NASA JSC has initiated an effort to improve and update the ORDEM96 model (Kessler et al. 1996, NASA TM-104825) utilizing the recently available data. This paper gives an overview of the new NASA orbital debris engineering model, ORDEM2000.

  13. The Solid Rocket Motor Slag Population: Results of a Radar-based Regressive Statistical Evaluation

    NASA Technical Reports Server (NTRS)

    Horstman, Matthew F.; Xu, Yu-Lin

    2008-01-01

    Solid rocket motor (SRM) slag has been identified as a significant source of man-made orbital debris. The propensity of SRMs to generate particles of 100 m and larger has caused concern regarding their contribution to the debris environment. Radar observation, rather than in-situ gathered evidence, is currently the only measurable source for the NASA/ODPO model of the on-orbit slag population. This simulated model includes the time evolution of the resultant orbital populations using a historical database of SRM launches, propellant masses, and estimated locations and times of tail-off. However, due to the small amount of observational evidence, there can be no direct comparison to check the validity of this model. Rather than using the assumed population developed from purely historical and physical assumptions, a regressional approach was used which utilized the populations observed by the Haystack radar from 1996 to present. The estimated trajectories from the historical model of slag sources, and the corresponding plausible detections by the Haystack radar, were identified. Comparisons with observational data from the ensuing years were made, and the SRM model was altered with respect to size and mass production of slag particles to reflect the historical data obtained. The result is a model SRM population that fits within the bounds of the observed environment.

  14. Identification and Functional Prediction of Large Intergenic Noncoding RNAs (lincRNAs) in Rainbow Trout (Oncorhynchus mykiss)

    USDA-ARS?s Scientific Manuscript database

    Long noncoding RNAs (lncRNAs) have been recognized in recent years as key regulators of diverse cellular processes. Genome-wide large-scale projects have uncovered thousands of lncRNAs in many model organisms. Large intergenic noncoding RNAs (lincRNAs) are lncRNAs that are transcribed from intergeni...

  15. The large-scale distribution of galaxies

    NASA Technical Reports Server (NTRS)

    Geller, Margaret J.

    1989-01-01

    The spatial distribution of galaxies in the universe is characterized on the basis of the six completed strips of the Harvard-Smithsonian Center for Astrophysics redshift-survey extension. The design of the survey is briefly reviewed, and the results are presented graphically. Vast low-density voids similar to the void in Bootes are found, almost completely surrounded by thin sheets of galaxies. Also discussed are the implications of the results for the survey sampling problem, the two-point correlation function of the galaxy distribution, the possibility of detecting large-scale coherent flows, theoretical models of large-scale structure, and the identification of groups and clusters of galaxies.

  16. Identification of Phosphorylated Proteins on a Global Scale.

    PubMed

    Iliuk, Anton

    2018-05-31

    Liquid chromatography (LC) coupled with tandem mass spectrometry (MS/MS) has enabled researchers to analyze complex biological samples with unprecedented depth. It facilitates the identification and quantification of modifications within thousands of proteins in a single large-scale proteomic experiment. Analysis of phosphorylation, one of the most common and important post-translational modifications, has particularly benefited from such progress in the field. Here, detailed protocols are provided for a few well-regarded, common sample preparation methods for an effective phosphoproteomic experiment. © 2018 by John Wiley & Sons, Inc. Copyright © 2018 John Wiley & Sons, Inc.

  17. Large-scale identification of target proteins of a glycosyltransferase isozyme by Lectin-IGOT-LC/MS, an LC/MS-based glycoproteomic approach

    PubMed Central

    Sugahara, Daisuke; Kaji, Hiroyuki; Sugihara, Kazushi; Asano, Masahide; Narimatsu, Hisashi

    2012-01-01

    Model organisms containing deletion or mutation in a glycosyltransferase-gene exhibit various physiological abnormalities, suggesting that specific glycan motifs on certain proteins play important roles in vivo. Identification of the target proteins of glycosyltransferase isozymes is the key to understand the roles of glycans. Here, we demonstrated the proteome-scale identification of the target proteins specific for a glycosyltransferase isozyme, β1,4-galactosyltransferase-I (β4GalT-I). Although β4GalT-I is the most characterized glycosyltransferase, its distinctive contribution to β1,4-galactosylation has been hardly described so far. We identified a large number of candidates for the target proteins specific to β4GalT-I by comparative analysis of β4GalT-I-deleted and wild-type mice using the LC/MS-based technique with the isotope-coded glycosylation site-specific tagging (IGOT) of lectin-captured N-glycopeptides. Our approach to identify the target proteins in a proteome-scale offers common features and trends in the target proteins, which facilitate understanding of the mechanism that controls assembly of a particular glycan motif on specific proteins. PMID:23002422

  18. Detection and classification of ash dieback on large-scale color aerial photographs

    Treesearch

    Ralph J. Croxton

    1966-01-01

    Aerial color photographs were taken at two scales over ash stands in New York State that were infected with ash dieback. Three photo interpreters then attempted to distinguish ash trees from other hardwoods and classify their disease condition. The scale of 1:7,920 was too small to permit accurate identification, but accuracy at the scale 1:1,584 was fair (60 to 70...

  19. Grant Development for Large Scale Research Proposals: An Overview and Case Study

    ERIC Educational Resources Information Center

    Goodman, Ira S.

    2011-01-01

    With some NIH pay lines running at or below the 10th percentile, and funding becoming scarce for large science grants, new approaches are necessary to secure large interdisciplinary grant awards. The UCSD Moores Cancer Center has developed a team approach, starting with the identification of a competitive opportunity and progressing to the…

  20. Genetics of Resistant Hypertension: the Missing Heritability and Opportunities.

    PubMed

    Teixeira, Samantha K; Pereira, Alexandre C; Krieger, Jose E

    2018-05-19

    Blood pressure regulation in humans has long been known to be a genetically determined trait. The identification of causal genetic modulators for this trait has been unfulfilling at the least. Despite the recent advances of genome-wide genetic studies, loci associated with hypertension or blood pressure still explain a very low percentage of the overall variation of blood pressure in the general population. This has precluded the translation of discoveries in the genetics of human hypertension to clinical use. Here, we propose the combined use of resistant hypertension as a trait for mapping genetic determinants in humans and the integration of new large-scale technologies to approach in model systems the multidimensional nature of the problem. New large-scale efforts in the genetic and genomic arenas are paving the way for an increased and granular understanding of genetic determinants of hypertension. New technologies for whole genome sequence and large-scale forward genetic screens can help prioritize gene and gene-pathways for downstream characterization and large-scale population studies, and guided pharmacological design can be used to drive discoveries to the translational application through better risk stratification and new therapeutic approaches. Although significant challenges remain in the mapping and identification of genetic determinants of hypertension, new large-scale technological approaches have been proposed to surpass some of the shortcomings that have limited progress in the area for the last three decades. The incorporation of these technologies to hypertension research may significantly help in the understanding of inter-individual blood pressure variation and the deployment of new phenotyping and treatment approaches for the condition.

  1. Exploring the performance of large-N radio astronomical arrays

    NASA Astrophysics Data System (ADS)

    Lonsdale, Colin J.; Doeleman, Sheperd S.; Cappallo, Roger J.; Hewitt, Jacqueline N.; Whitney, Alan R.

    2000-07-01

    New radio telescope arrays are currently being contemplated which may be built using hundreds, or even thousands, of relatively small antennas. These include the One Hectare Telescope of the SETI Institute and UC Berkeley, the LOFAR telescope planned for the New Mexico desert surrounding the VLA, and possibly the ambitious international Square Kilometer Array (SKA) project. Recent and continuing advances in signal transmission and processing technology make it realistic to consider full cross-correlation of signals from such a large number of antennas, permitting the synthesis of an aperture with much greater fidelity than in the past. In principle, many advantages in instrumental performance are gained by this 'large-N' approach to the design, most of which require the development of new algorithms. Because new instruments of this type are expected to outstrip the performance of current instruments by wide margins, much of their scientific productivity is likely to come from the study of objects which are currently unknown. For this reason, instrumental flexibility is of special importance in design studies. A research effort has begun at Haystack Observatory to explore large-N performance benefits, and to determine what array design properties and data reduction algorithms are required to achieve them. The approach to these problems, involving a sophisticated data simulator, algorithm development, and exploration of array configuration parameter space, will be described, and progress to date will be summarized.

  2. Precise excision and self-integration of a composite transposon as a model for spontaneous large-scale chromosome inversion/deletion of the Staphylococcus haemolyticus clinical strain JCSC1435.

    PubMed

    Watanabe, Shinya; Ito, Teruyo; Morimoto, Yuh; Takeuchi, Fumihiko; Hiramatsu, Keiichi

    2007-04-01

    Large-scale chromosomal inversions (455 to 535 kbp) or deletions (266 to 320 kbp) were found to accompany spontaneous loss of beta-lactam resistance during drug-free passage of the multiresistant Staphylococcus haemolyticus clinical strain JCSC1435. Identification and sequencing of the rearranged chromosomal loci revealed that ISSha1 of S. haemolyticus is responsible for the chromosome rearrangements.

  3. A global traveling wave on Venus

    NASA Technical Reports Server (NTRS)

    Smith, Michael D.; Gierasch, Peter J.; Schinder, Paul J.

    1993-01-01

    The dominant large-scale pattern in the clouds of Venus has been described as a 'Y' or 'Psi' and tentatively identified by earlier workers as a Kelvin wave. A detailed calculation of linear wave modes in the Venus atmosphere verifies this identification. Cloud feedback by infrared heating fluctuations is a plausible excitation mechanism. Modulation of the large-scale pattern by the wave is a possible explanation for the Y. Momentum transfer by the wave could contribute to sustaining the general circulation.

  4. Towards Development of Clustering Applications for Large-Scale Comparative Genotyping and Kinship Analysis Using Y-Short Tandem Repeats.

    PubMed

    Seman, Ali; Sapawi, Azizian Mohd; Salleh, Mohd Zaki

    2015-06-01

    Y-chromosome short tandem repeats (Y-STRs) are genetic markers with practical applications in human identification. However, where mass identification is required (e.g., in the aftermath of disasters with significant fatalities), the efficiency of the process could be improved with new statistical approaches. Clustering applications are relatively new tools for large-scale comparative genotyping, and the k-Approximate Modal Haplotype (k-AMH), an efficient algorithm for clustering large-scale Y-STR data, represents a promising method for developing these tools. In this study we improved the k-AMH and produced three new algorithms: the Nk-AMH I (including a new initial cluster center selection), the Nk-AMH II (including a new dominant weighting value), and the Nk-AMH III (combining I and II). The Nk-AMH III was the superior algorithm, with mean clustering accuracy that increased in four out of six datasets and remained at 100% in the other two. Additionally, the Nk-AMH III achieved a 2% higher overall mean clustering accuracy score than the k-AMH, as well as optimal accuracy for all datasets (0.84-1.00). With inclusion of the two new methods, the Nk-AMH III produced an optimal solution for clustering Y-STR data; thus, the algorithm has potential for further development towards fully automatic clustering of any large-scale genotypic data.

  5. Parameter identification of civil engineering structures

    NASA Technical Reports Server (NTRS)

    Juang, J. N.; Sun, C. T.

    1980-01-01

    This paper concerns the development of an identification method required in determining structural parameter variations for systems subjected to an extended exposure to the environment. The concept of structural identifiability of a large scale structural system in the absence of damping is presented. Three criteria are established indicating that a large number of system parameters (the coefficient parameters of the differential equations) can be identified by a few actuators and sensors. An eight-bay-fifteen-story frame structure is used as example. A simple model is employed for analyzing the dynamic response of the frame structure.

  6. How Should Witnesses Regulate the Accuracy of Their Identification Decisions: One Step Forward, Two Steps Back?

    ERIC Educational Resources Information Center

    Perfect, Timothy J.; Weber, Nathan

    2012-01-01

    Explorations of memory accuracy control normally contrast forced-report with free-report performance across a set of items and show a trade-off between memory quantity and accuracy. However, this memory control framework has not been tested with lineup identifications that may involve rejection of all alternatives. A large-scale (N = 439) lineup…

  7. Groups of galaxies in the Center for Astrophysics redshift survey

    NASA Technical Reports Server (NTRS)

    Ramella, Massimo; Geller, Margaret J.; Huchra, John P.

    1989-01-01

    By applying the Huchra and Geller (1982) objective group identification algorithm to the Center for Astrophysics' redshift survey, a catalog of 128 groups with three or more members is extracted, and 92 of these are used as a statistical sample. A comparison of the distribution of group centers with the distribution of all galaxies in the survey indicates qualitatively that groups trace the large-scale structure of the region. The physical properties of groups may be related to the details of large-scale structure, and it is concluded that differences among group catalogs may be due to the properties of large-scale structures and their location relative to the survey limits.

  8. Chemocoding as an identification tool where morphological- and DNA-based methods fall short: Inga as a case study.

    PubMed

    Endara, María-José; Coley, Phyllis D; Wiggins, Natasha L; Forrister, Dale L; Younkin, Gordon C; Nicholls, James A; Pennington, R Toby; Dexter, Kyle G; Kidner, Catherine A; Stone, Graham N; Kursar, Thomas A

    2018-04-01

    The need for species identification and taxonomic discovery has led to the development of innovative technologies for large-scale plant identification. DNA barcoding has been useful, but fails to distinguish among many species in species-rich plant genera, particularly in tropical regions. Here, we show that chemical fingerprinting, or 'chemocoding', has great potential for plant identification in challenging tropical biomes. Using untargeted metabolomics in combination with multivariate analysis, we constructed species-level fingerprints, which we define as chemocoding. We evaluated the utility of chemocoding with species that were defined morphologically and subject to next-generation DNA sequencing in the diverse and recently radiated neotropical genus Inga (Leguminosae), both at single study sites and across broad geographic scales. Our results show that chemocoding is a robust method for distinguishing morphologically similar species at a single site and for identifying widespread species across continental-scale ranges. Given that species are the fundamental unit of analysis for conservation and biodiversity research, the development of accurate identification methods is essential. We suggest that chemocoding will be a valuable additional source of data for a quick identification of plants, especially for groups where other methods fall short. © 2018 The Authors. New Phytologist © 2018 New Phytologist Trust.

  9. A needle in a haystack? Uninsured workers in small businesses that do not offer coverage.

    PubMed

    Kronick, Richard; Olsen, Louis C

    2006-02-01

    To describe the insurance status of workers at small businesses, and to describe the status of uninsured persons by the employment characteristics (employment status, firm size, and whether the employer offers insurance) of the head of household. Data from the March and February 2001 Current Population Survey, and a survey of 2,830 small businesses in San Diego County conducted in 2001. The survey of small businesses was undertaken as part of a project testing the response of employers to offers of subsidized coverage. Employers were asked whether they offered insurance, and about the insurance status of their employees. The merged February-March 2001 CPS was used to identify the employment status, firm size, and employer-offering status for uninsured persons in the U.S. Telephone interviews with small businesses in San Diego County. Only 21 percent of the uninsured in the U.S. are full-time employees (or their dependents) in small businesses (<100 employees) that do not offer insurance. The employment status of the uninsured is heterogeneous: many work for large employers, small employers who do offer insurance, or are self-employed, part-time workers, or have no workers in the household. Although there are many small businesses in San Diego that do not offer coverage, most of them have very few uninsured workers. Over 50 percent of businesses that do not offer coverage have either zero or one uninsured worker. There are very few small businesses that do not offer coverage and that have substantial numbers of uninsured workers. These businesses are not quite as rare as a needle in a haystack, but they are very difficult to find. If all small businesses that do not offer insurance now could be persuaded to start offering coverage, and if all the full-time workers (and their dependents) in those businesses accepted insurance, the number of uninsured would decline by 21 percent--a significant decline, but leaving 80 percent of the problem untouched. If the prime target for programs of subsidized insurance are small businesses that do not offer coverage now and that have substantial numbers of uninsured workers, the target is very small.

  10. Hit discovery of 4-amino-N-(4-(3-(trifluoromethyl)phenoxy)pyrimidin-5-yl)benzamide: A novel EGFR inhibitor from a designed small library.

    PubMed

    Elkamhawy, Ahmed; Paik, Sora; Hassan, Ahmed H E; Lee, Yong Sup; Roh, Eun Joo

    2017-12-01

    Searching for hit compounds within the huge chemical space resembles the attempt to find a needle in a haystack. Cheminformatics-guided selection of few representative molecules of a rationally designed virtual combinatorial library is a powerful tool to confront this challenge, speed up hit identification and cut off costs. Herein, this approach has been applied to identify hit compounds with novel scaffolds able to inhibit EGFR kinase. From a generated virtual library, six 4-aryloxy-5-aminopyrimidine scaffold-derived compounds were selected, synthesized and evaluated as hit EGFR inhibitors. 4-Aryloxy-5-benzamidopyrimidines inhibited EGFR with IC 50 1.05-5.37 μM. Cell-based assay of the most potent EGFR inhibitor hit (10ac) confirmed its cytotoxicity against different cancerous cells. In spite of no EGFR, HER2 or VEGFR1 inhibition was elicited by 4-aryloxy-5-(thio)ureidopyrimidine derivatives, cell-based evaluation suggested them as antiproliferative hits acting by other mechanism(s). Molecular docking study provided a plausible explanation of incapability of 4-aryloxy-5-(thio)ureidopyrimidines to inhibit EGFR and suggested a reasonable binding mode of 4-aryloxy-5-benzamidopyrimidines which provides a basis to develop more optimized ligands. Copyright © 2017 Elsevier Inc. All rights reserved.

  11. Haystack Antenna Control System Design Document

    DTIC Science & Technology

    2010-12-07

    The ICDs will be referenced, where appropriate. The control system isn’t being designed in a vacuum . Other teams are designing or will be designing...a horizontally scrolling display which updates in real time based upon instrumentation status messages from the ACU. In the above figure a rather...hydrostatic bearing pump systems are shut down. 6.8 ELEVATION STOW PINS The stow pins will be monitored and controlled via the PLC. There will be 2 or 4

  12. High Performance Visualization using Query-Driven Visualizationand Analytics

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bethel, E. Wes; Campbell, Scott; Dart, Eli

    2006-06-15

    Query-driven visualization and analytics is a unique approach for high-performance visualization that offers new capabilities for knowledge discovery and hypothesis testing. The new capabilities akin to finding needles in haystacks are the result of combining technologies from the fields of scientific visualization and scientific data management. This approach is crucial for rapid data analysis and visualization in the petascale regime. This article describes how query-driven visualization is applied to a hero-sized network traffic analysis problem.

  13. Finding a Single Molecule in a Haystack: Optical Detection and Spectroscopy of Single Absorbers in Solids

    DTIC Science & Technology

    1989-08-18

    CODES 18 SUBJECT TERMS (Continue on reverse if necessary and identify by block number) FIELD GROUP SUB-GROUP Single Molecule Detection Pentacene in p...and 10 additional pentacene molecules. This may be accomplished by- a combination of laser FM spectroscopy and either Stark or ultrasonic double...6099 408-927-2426 ABSTRACT: Single-absorber optical spectroscopy in solids is described for the case of finding a single pentacene molecule in a

  14. Army Logistician. Volume 35, Issue 4, July-August 2003

    DTIC Science & Technology

    2003-08-01

    DPOCs ). MPOCs and DPOCs are mobile MTMC offices that provide the same infor- mation technology capabilities that MTMC personnel have at their home...secondary ports, and DPOCs are shelterized and are used for major or regional conflicts. MPOCs and DPOCs provide MTMC with C4 systems to control and...finding items in the maze of pallets and con- tainers was like searching for the proverbial needle in a haystack. Now, thanks to MPOCs, DPOCs , MMCS, and

  15. Energy Innovation Portal Brings DOE Technologies to the Market (Fact Sheet)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    2011-10-01

    For venture capitalists, energy entrepreneurs, and industry veterans, finding the right renewable energy or energy efficiency solution used to be like looking for a needle in a haystack. Now, a searchable treasure trove of innovative U.S. Department of Energy (DOE) technologies is available. Created by the National Renewable Energy Laboratory (NREL), the online Energy Innovation Portal helps businesses and entrepreneurs access the intellectual property of DOE's 17 national laboratories and other research partners.

  16. Pharmaceutical Raw Material Identification Using Miniature Near-Infrared (MicroNIR) Spectroscopy and Supervised Pattern Recognition Using Support Vector Machine

    PubMed Central

    Hsiung, Chang; Pederson, Christopher G.; Zou, Peng; Smith, Valton; von Gunten, Marc; O’Brien, Nada A.

    2016-01-01

    Near-infrared spectroscopy as a rapid and non-destructive analytical technique offers great advantages for pharmaceutical raw material identification (RMID) to fulfill the quality and safety requirements in pharmaceutical industry. In this study, we demonstrated the use of portable miniature near-infrared (MicroNIR) spectrometers for NIR-based pharmaceutical RMID and solved two challenges in this area, model transferability and large-scale classification, with the aid of support vector machine (SVM) modeling. We used a set of 19 pharmaceutical compounds including various active pharmaceutical ingredients (APIs) and excipients and six MicroNIR spectrometers to test model transferability. For the test of large-scale classification, we used another set of 253 pharmaceutical compounds comprised of both chemically and physically different APIs and excipients. We compared SVM with conventional chemometric modeling techniques, including soft independent modeling of class analogy, partial least squares discriminant analysis, linear discriminant analysis, and quadratic discriminant analysis. Support vector machine modeling using a linear kernel, especially when combined with a hierarchical scheme, exhibited excellent performance in both model transferability and large-scale classification. Hence, ultra-compact, portable and robust MicroNIR spectrometers coupled with SVM modeling can make on-site and in situ pharmaceutical RMID for large-volume applications highly achievable. PMID:27029624

  17. A workflow for large-scale empirical identification of cell wall N-linked glycoproteins of tomato (Solanum lycopersicum) fruit by tandem mass spectrometry

    USDA-ARS?s Scientific Manuscript database

    Glycosylation is a common post-translational modification of plant proteins that impacts a large number of important biological processes. Nevertheless, the impacts of differential site occupancy and the nature of specific glycoforms are obscure. Historically, characterization of glycoproteins has b...

  18. Aerial photo guide to New England forest cover types

    Treesearch

    Rachel Riemann Hershey; William A. Befort

    1995-01-01

    NOTE large file size. Presents color infrared photos in stereo pairs for the identification of New England forest cover types. Depicts range maps, ecological relations, and range of composition for each forest cover type described. The guide is designed to assist the needs of interpreters of medium to large-scale color infrared aerial photography.

  19. Resources for Functional Genomics Studies in Drosophila melanogaster

    PubMed Central

    Mohr, Stephanie E.; Hu, Yanhui; Kim, Kevin; Housden, Benjamin E.; Perrimon, Norbert

    2014-01-01

    Drosophila melanogaster has become a system of choice for functional genomic studies. Many resources, including online databases and software tools, are now available to support design or identification of relevant fly stocks and reagents or analysis and mining of existing functional genomic, transcriptomic, proteomic, etc. datasets. These include large community collections of fly stocks and plasmid clones, “meta” information sites like FlyBase and FlyMine, and an increasing number of more specialized reagents, databases, and online tools. Here, we introduce key resources useful to plan large-scale functional genomics studies in Drosophila and to analyze, integrate, and mine the results of those studies in ways that facilitate identification of highest-confidence results and generation of new hypotheses. We also discuss ways in which existing resources can be used and might be improved and suggest a few areas of future development that would further support large- and small-scale studies in Drosophila and facilitate use of Drosophila information by the research community more generally. PMID:24653003

  20. Musical expertise is related to altered functional connectivity during audiovisual integration

    PubMed Central

    Paraskevopoulos, Evangelos; Kraneburg, Anja; Herholz, Sibylle Cornelia; Bamidis, Panagiotis D.; Pantev, Christo

    2015-01-01

    The present study investigated the cortical large-scale functional network underpinning audiovisual integration via magnetoencephalographic recordings. The reorganization of this network related to long-term musical training was investigated by comparing musicians to nonmusicians. Connectivity was calculated on the basis of the estimated mutual information of the sources’ activity, and the corresponding networks were statistically compared. Nonmusicians’ results indicated that the cortical network associated with audiovisual integration supports visuospatial processing and attentional shifting, whereas a sparser network, related to spatial awareness supports the identification of audiovisual incongruences. In contrast, musicians’ results showed enhanced connectivity in regions related to the identification of auditory pattern violations. Hence, nonmusicians rely on the processing of visual clues for the integration of audiovisual information, whereas musicians rely mostly on the corresponding auditory information. The large-scale cortical network underpinning multisensory integration is reorganized due to expertise in a cognitive domain that largely involves audiovisual integration, indicating long-term training-related neuroplasticity. PMID:26371305

  1. Large-Scale Chemical Similarity Networks for Target Profiling of Compounds Identified in Cell-Based Chemical Screens

    PubMed Central

    Lo, Yu-Chen; Senese, Silvia; Li, Chien-Ming; Hu, Qiyang; Huang, Yong; Damoiseaux, Robert; Torres, Jorge Z.

    2015-01-01

    Target identification is one of the most critical steps following cell-based phenotypic chemical screens aimed at identifying compounds with potential uses in cell biology and for developing novel disease therapies. Current in silico target identification methods, including chemical similarity database searches, are limited to single or sequential ligand analysis that have limited capabilities for accurate deconvolution of a large number of compounds with diverse chemical structures. Here, we present CSNAP (Chemical Similarity Network Analysis Pulldown), a new computational target identification method that utilizes chemical similarity networks for large-scale chemotype (consensus chemical pattern) recognition and drug target profiling. Our benchmark study showed that CSNAP can achieve an overall higher accuracy (>80%) of target prediction with respect to representative chemotypes in large (>200) compound sets, in comparison to the SEA approach (60–70%). Additionally, CSNAP is capable of integrating with biological knowledge-based databases (Uniprot, GO) and high-throughput biology platforms (proteomic, genetic, etc) for system-wise drug target validation. To demonstrate the utility of the CSNAP approach, we combined CSNAP's target prediction with experimental ligand evaluation to identify the major mitotic targets of hit compounds from a cell-based chemical screen and we highlight novel compounds targeting microtubules, an important cancer therapeutic target. The CSNAP method is freely available and can be accessed from the CSNAP web server (http://services.mbi.ucla.edu/CSNAP/). PMID:25826798

  2. Classification of Large-Scale Remote Sensing Images for Automatic Identification of Health Hazards: Smoke Detection Using an Autologistic Regression Classifier.

    PubMed

    Wolters, Mark A; Dean, C B

    2017-01-01

    Remote sensing images from Earth-orbiting satellites are a potentially rich data source for monitoring and cataloguing atmospheric health hazards that cover large geographic regions. A method is proposed for classifying such images into hazard and nonhazard regions using the autologistic regression model, which may be viewed as a spatial extension of logistic regression. The method includes a novel and simple approach to parameter estimation that makes it well suited to handling the large and high-dimensional datasets arising from satellite-borne instruments. The methodology is demonstrated on both simulated images and a real application to the identification of forest fire smoke.

  3. Searching for the elusive gift: advances in talent identification in sport.

    PubMed

    Mann, David L; Dehghansai, Nima; Baker, Joseph

    2017-08-01

    The incentives for sport organizations to identify talented athletes from a young age continue to grow, yet effective talent identification remains a challenging task. This opinion paper examines recent advances in talent identification, focusing in particular on the emergence of new approaches that may offer promise to identify talent (e.g., small-sided games, genetic testing, and advanced statistical analyses). We appraise new multi-disciplinary and large-scale population studies of talent identification, provide a consideration of the most recent psychological predictors of performance, examine the emergence of new approaches that strive to diminish biases in talent identification, and look at the rise in interest in talent identification in Paralympic sport. Crown Copyright © 2017. Published by Elsevier Ltd. All rights reserved.

  4. Regression-Based Identification of Behavior-Encoding Neurons During Large-Scale Optical Imaging of Neural Activity at Cellular Resolution

    PubMed Central

    Miri, Andrew; Daie, Kayvon; Burdine, Rebecca D.; Aksay, Emre

    2011-01-01

    The advent of methods for optical imaging of large-scale neural activity at cellular resolution in behaving animals presents the problem of identifying behavior-encoding cells within the resulting image time series. Rapid and precise identification of cells with particular neural encoding would facilitate targeted activity measurements and perturbations useful in characterizing the operating principles of neural circuits. Here we report a regression-based approach to semiautomatically identify neurons that is based on the correlation of fluorescence time series with quantitative measurements of behavior. The approach is illustrated with a novel preparation allowing synchronous eye tracking and two-photon laser scanning fluorescence imaging of calcium changes in populations of hindbrain neurons during spontaneous eye movement in the larval zebrafish. Putative velocity-to-position oculomotor integrator neurons were identified that showed a broad spatial distribution and diversity of encoding. Optical identification of integrator neurons was confirmed with targeted loose-patch electrical recording and laser ablation. The general regression-based approach we demonstrate should be widely applicable to calcium imaging time series in behaving animals. PMID:21084686

  5. [Organizational problems of disaster victim identification in mass casualties as exemplified by Tu 154-M and Airbus A310 passenger plane crashes].

    PubMed

    Volkov, A V; Kolkutin, V V; Klevno, V A; Shkol'nikov, B V; Kornienko, I V

    2008-01-01

    Managerial experience is described that was gained during the large-scale work on victim identification following mass casualties in the Tu 154-M and Airbus A310 passenger plane crashes. The authors emphasize the necessity to set up a specialized agency of constant readiness meeting modern requirements for the implementation of a system of measures for personality identification. This agency must incorporate relevant departments of the Ministries of Health, Defense, and Emergency Situations as well as investigative authorities and other organizations.

  6. Large Scale Single Nucleotide Polymorphism Study of PD Susceptibility

    DTIC Science & Technology

    2005-03-01

    identification of eight genetic loci in the familial PD, the results of intensive investigations of polymorphisms in dozens of genes related to sporadic, late...1) investigate the association between classical, sporadic PD and 2386 SNPs in 23 genes implicated in the pathogenesis of PD; (2) construct...addition, experiences derived from this study may be applied in other complex disorders for the identification of susceptibility genes , as well as in genome

  7. Rapid group-, serotype-, and vaccine strain-specific identification of poliovirus isolates by real-time reverse transcription-PCR using degenerate primers and probes containing deoxyinosine residues.

    PubMed

    Kilpatrick, David R; Yang, Chen-Fu; Ching, Karen; Vincent, Annelet; Iber, Jane; Campagnoli, Ray; Mandelbaum, Mark; De, Lina; Yang, Su-Ju; Nix, Allan; Kew, Olen M

    2009-06-01

    We have adapted our previously described poliovirus diagnostic reverse transcription-PCR (RT-PCR) assays to a real-time RT-PCR (rRT-PCR) format. Our highly specific assays and rRT-PCR reagents are designed for use in the WHO Global Polio Laboratory Network for rapid and large-scale identification of poliovirus field isolates.

  8. Crustal deformation at very long baseline interferometry sites due to seasonal air-mass and ground water variations

    NASA Technical Reports Server (NTRS)

    Stolz, A.; Larden, D. R.

    1980-01-01

    The seasonal deformation normal to the Earth's surface was calculated at stations involved or interested in very long baseline interferometry (VLBI) geodesy and at hypothetical sites in Australia and Brazil using global atmospheric pressure data, values for groundwater storage, and load Love numbers deduced from current Earth models. It was found that the annual range of deformation approached the centimeter level measuring potential of the VLBI technqiue at Greenbank, Haystack, and the Brazil site.

  9. Vector Antenna and Maximum Likelihood Imaging for Radio Astronomy

    DTIC Science & Technology

    2016-03-05

    Maximum Likelihood Imaging for Radio Astronomy Mary Knapp1, Frank Robey2, Ryan Volz3, Frank Lind3, Alan Fenn2, Alex Morris2, Mark Silver2, Sarah Klein2...haystack.mit.edu Abstract1— Radio astronomy using frequencies less than ~100 MHz provides a window into non-thermal processes in objects ranging from planets...observational astronomy . Ground-based observatories including LOFAR [1], LWA [2], [3], MWA [4], and the proposed SKA-Low [5], [6] are improving access to

  10. VLBI Digital-Backend Intercomparison Test Report

    NASA Technical Reports Server (NTRS)

    Whitney, Alan; Beaudoin, Christopher; Cappallo, Roger; Niell, Arthur; Petrachenko, Bill; Ruszczyk, Chester A.; Titus, Mike

    2013-01-01

    Issues related to digital-backend (DBE) systems can be difficult to evaluate in either local tests or actual VLBI experiments. The 2nd DBE intercomparison workshop at Haystack Observatory on 25-26 October 2012 provided a forum to explicitly address validation and interoperability issues among independent global developers of DBE equipment. This special report discusses the workshop. It identifies DBE systems that were tested at the workshop, describes the test objectives and procedures, and reports and discusses the results of the testing.

  11. Finding a Single Molecule in a Haystack: Laser Spectroscopy of Solids from Sqrt. N to N = 1

    DTIC Science & Technology

    1991-01-02

    low-temperature inhomogeneously broadened 0-0 S, +- S electronic transition of pentacene dopant molecules in p-terphenyl crystals have yielded both (1...absorber, single-miolecule detection, or SMDI) wvould provide a-- useful tool for the study of local host-absorber interactions wiherte the absorbing...molecular impurity is-a truly local probe of the minute details of a single local environment in a solid. l’he use of powerful spectroscopic methods as

  12. A Review of Challenges in Developing a National Program for Gifted Children in India's Diverse Context

    ERIC Educational Resources Information Center

    Kurup, Anitha; Maithreyi, R.

    2012-01-01

    Large-scale sequential research developments for identification and measurement of giftedness have received ample attention in the West, whereas India's response to this has largely been lukewarm. The wide variation in parents' abilities to provide enriched environments to nurture their children's potential makes it imperative for India to develop…

  13. Advancing the large-scale CCS database for metabolomics and lipidomics at the machine-learning era.

    PubMed

    Zhou, Zhiwei; Tu, Jia; Zhu, Zheng-Jiang

    2018-02-01

    Metabolomics and lipidomics aim to comprehensively measure the dynamic changes of all metabolites and lipids that are present in biological systems. The use of ion mobility-mass spectrometry (IM-MS) for metabolomics and lipidomics has facilitated the separation and the identification of metabolites and lipids in complex biological samples. The collision cross-section (CCS) value derived from IM-MS is a valuable physiochemical property for the unambiguous identification of metabolites and lipids. However, CCS values obtained from experimental measurement and computational modeling are limited available, which significantly restricts the application of IM-MS. In this review, we will discuss the recently developed machine-learning based prediction approach, which could efficiently generate precise CCS databases in a large scale. We will also highlight the applications of CCS databases to support metabolomics and lipidomics. Copyright © 2017 Elsevier Ltd. All rights reserved.

  14. Exhaustive identification of steady state cycles in large stoichiometric networks

    PubMed Central

    Wright, Jeremiah; Wagner, Andreas

    2008-01-01

    Background Identifying cyclic pathways in chemical reaction networks is important, because such cycles may indicate in silico violation of energy conservation, or the existence of feedback in vivo. Unfortunately, our ability to identify cycles in stoichiometric networks, such as signal transduction and genome-scale metabolic networks, has been hampered by the computational complexity of the methods currently used. Results We describe a new algorithm for the identification of cycles in stoichiometric networks, and we compare its performance to two others by exhaustively identifying the cycles contained in the genome-scale metabolic networks of H. pylori, M. barkeri, E. coli, and S. cerevisiae. Our algorithm can substantially decrease both the execution time and maximum memory usage in comparison to the two previous algorithms. Conclusion The algorithm we describe improves our ability to study large, real-world, biochemical reaction networks, although additional methodological improvements are desirable. PMID:18616835

  15. Computational Issues in Damping Identification for Large Scale Problems

    NASA Technical Reports Server (NTRS)

    Pilkey, Deborah L.; Roe, Kevin P.; Inman, Daniel J.

    1997-01-01

    Two damping identification methods are tested for efficiency in large-scale applications. One is an iterative routine, and the other a least squares method. Numerical simulations have been performed on multiple degree-of-freedom models to test the effectiveness of the algorithm and the usefulness of parallel computation for the problems. High Performance Fortran is used to parallelize the algorithm. Tests were performed using the IBM-SP2 at NASA Ames Research Center. The least squares method tested incurs high communication costs, which reduces the benefit of high performance computing. This method's memory requirement grows at a very rapid rate meaning that larger problems can quickly exceed available computer memory. The iterative method's memory requirement grows at a much slower pace and is able to handle problems with 500+ degrees of freedom on a single processor. This method benefits from parallelization, and significant speedup can he seen for problems of 100+ degrees-of-freedom.

  16. Tools for phospho- and glycoproteomics of plasma membranes.

    PubMed

    Wiśniewski, Jacek R

    2011-07-01

    Analysis of plasma membrane proteins and their posttranslational modifications is considered as important for identification of disease markers and targets for drug treatment. Due to their insolubility in water, studying of plasma membrane proteins using mass spectrometry has been difficult for a long time. Recent technological developments in sample preparation together with important improvements in mass spectrometric analysis have facilitated analysis of these proteins and their posttranslational modifications. Now, large scale proteomic analyses allow identification of thousands of membrane proteins from minute amounts of sample. Optimized protocols for affinity enrichment of phosphorylated and glycosylated peptides have set new dimensions in the depth of characterization of these posttranslational modifications of plasma membrane proteins. Here, I summarize recent advances in proteomic technology for the characterization of the cell surface proteins and their modifications. In the focus are approaches allowing large scale mapping rather than analytical methods suitable for studying individual proteins or non-complex mixtures.

  17. Accurate population genetic measurements require cryptic species identification in corals

    NASA Astrophysics Data System (ADS)

    Sheets, Elizabeth A.; Warner, Patricia A.; Palumbi, Stephen R.

    2018-06-01

    Correct identification of closely related species is important for reliable measures of gene flow. Incorrectly lumping individuals of different species together has been shown to over- or underestimate population differentiation, but examples highlighting when these different results are observed in empirical datasets are rare. Using 199 single nucleotide polymorphisms, we assigned 768 individuals in the Acropora hyacinthus and A. cytherea morphospecies complexes to each of eight previously identified cryptic genetic species and measured intraspecific genetic differentiation across three geographic scales (within reefs, among reefs within an archipelago, and among Pacific archipelagos). We then compared these calculations to estimated genetic differentiation at each scale with all cryptic genetic species mixed as if we could not tell them apart. At the reef scale, correct genetic species identification yielded lower F ST estimates and fewer significant comparisons than when species were mixed, raising estimates of short-scale gene flow. In contrast, correct genetic species identification at large spatial scales yielded higher F ST measurements than mixed-species comparisons, lowering estimates of long-term gene flow among archipelagos. A meta-analysis of published population genetic studies in corals found similar results: F ST estimates at small spatial scales were lower and significance was found less often in studies that controlled for cryptic species. Our results and these prior datasets controlling for cryptic species suggest that genetic differentiation among local reefs may be lower than what has generally been reported in the literature. Not properly controlling for cryptic species structure can bias population genetic analyses in different directions across spatial scales, and this has important implications for conservation strategies that rely on these estimates.

  18. Preliminary investigation of Large Format Camera photography utility in soil mapping and related agricultural applications

    NASA Technical Reports Server (NTRS)

    Pelletier, R. E.; Hudnall, W. H.

    1987-01-01

    The use of Space Shuttle Large Format Camera (LFC) color, IR/color, and B&W images in large-scale soil mapping is discussed and illustrated with sample photographs from STS 41-6 (October 1984). Consideration is given to the characteristics of the film types used; the photographic scales available; geometric and stereoscopic factors; and image interpretation and classification for soil-type mapping (detecting both sharp and gradual boundaries), soil parent material topographic and hydrologic assessment, natural-resources inventory, crop-type identification, and stress analysis. It is suggested that LFC photography can play an important role, filling the gap between aerial and satellite remote sensing.

  19. Empirical performance of the self-controlled case series design: lessons for developing a risk identification and analysis system.

    PubMed

    Suchard, Marc A; Zorych, Ivan; Simpson, Shawn E; Schuemie, Martijn J; Ryan, Patrick B; Madigan, David

    2013-10-01

    The self-controlled case series (SCCS) offers potential as an statistical method for risk identification involving medical products from large-scale observational healthcare data. However, analytic design choices remain in encoding the longitudinal health records into the SCCS framework and its risk identification performance across real-world databases is unknown. To evaluate the performance of SCCS and its design choices as a tool for risk identification in observational healthcare data. We examined the risk identification performance of SCCS across five design choices using 399 drug-health outcome pairs in five real observational databases (four administrative claims and one electronic health records). In these databases, the pairs involve 165 positive controls and 234 negative controls. We also consider several synthetic databases with known relative risks between drug-outcome pairs. We evaluate risk identification performance through estimating the area under the receiver-operator characteristics curve (AUC) and bias and coverage probability in the synthetic examples. The SCCS achieves strong predictive performance. Twelve of the twenty health outcome-database scenarios return AUCs >0.75 across all drugs. Including all adverse events instead of just the first per patient and applying a multivariate adjustment for concomitant drug use are the most important design choices. However, the SCCS as applied here returns relative risk point-estimates biased towards the null value of 1 with low coverage probability. The SCCS recently extended to apply a multivariate adjustment for concomitant drug use offers promise as a statistical tool for risk identification in large-scale observational healthcare databases. Poor estimator calibration dampens enthusiasm, but on-going work should correct this short-coming.

  20. Development and in silico evaluation of large-scale metabolite identification methods using functional group detection for metabolomics

    PubMed Central

    Mitchell, Joshua M.; Fan, Teresa W.-M.; Lane, Andrew N.; Moseley, Hunter N. B.

    2014-01-01

    Large-scale identification of metabolites is key to elucidating and modeling metabolism at the systems level. Advances in metabolomics technologies, particularly ultra-high resolution mass spectrometry (MS) enable comprehensive and rapid analysis of metabolites. However, a significant barrier to meaningful data interpretation is the identification of a wide range of metabolites including unknowns and the determination of their role(s) in various metabolic networks. Chemoselective (CS) probes to tag metabolite functional groups combined with high mass accuracy provide additional structural constraints for metabolite identification and quantification. We have developed a novel algorithm, Chemically Aware Substructure Search (CASS) that efficiently detects functional groups within existing metabolite databases, allowing for combined molecular formula and functional group (from CS tagging) queries to aid in metabolite identification without a priori knowledge. Analysis of the isomeric compounds in both Human Metabolome Database (HMDB) and KEGG Ligand demonstrated a high percentage of isomeric molecular formulae (43 and 28%, respectively), indicating the necessity for techniques such as CS-tagging. Furthermore, these two databases have only moderate overlap in molecular formulae. Thus, it is prudent to use multiple databases in metabolite assignment, since each major metabolite database represents different portions of metabolism within the biosphere. In silico analysis of various CS-tagging strategies under different conditions for adduct formation demonstrate that combined FT-MS derived molecular formulae and CS-tagging can uniquely identify up to 71% of KEGG and 37% of the combined KEGG/HMDB database vs. 41 and 17%, respectively without adduct formation. This difference between database isomer disambiguation highlights the strength of CS-tagging for non-lipid metabolite identification. However, unique identification of complex lipids still needs additional information. PMID:25120557

  1. Mars synthetic topographic mapping

    USGS Publications Warehouse

    Wu, S.S.C.

    1978-01-01

    Topographic contour maps of Mars are compiled by the synthesis of data acquired from various scientific experiments of the Mariner 9 mission, including S-band radio-occulation, the ultraviolet spectrometer (UVS), the infrared radiometer (IRR), the infrared interferometer spectrometer (IRIS) and television imagery, as well as Earth-based radar information collected at Goldstone, Haystack, and Arecibo Observatories. The entire planet is mapped at scales of 1:25,000,000 and 1:25,000,000 using Mercator, Lambert, and polar stereographic map projections. For the computation of map projections, a biaxial spheroid figure is adopted. The semimajor and semiminor axes are 3393.4 and 3375.7 km, respectively, with a polar flattening of 0.0052. For the computation of elevations, a topographic datum is defined by a gravity field described in terms of spherical harmonics of fourth order and fourth degree combined with a 6.1-mbar occulation pressure surface. This areoid can be approximated by a triaxial ellipsoid with semimajor axes of A = 3394.6 km and B = 3393.3 km and a semiminor axis of C = 3376.3 km. The semimajor axis A intersects the Martian surface at longitude 105??W. The dynamic flattening of Mars is 0.00525. The contour intercal of the maps is 1 km. For some prominent features where overlapping pictures from Mariner 9 are available, local contour maps at relatively larger scales were also compiled by photogrammetric methods on stereo plotters. ?? 1978.

  2. ProteinInferencer: Confident protein identification and multiple experiment comparison for large scale proteomics projects.

    PubMed

    Zhang, Yaoyang; Xu, Tao; Shan, Bing; Hart, Jonathan; Aslanian, Aaron; Han, Xuemei; Zong, Nobel; Li, Haomin; Choi, Howard; Wang, Dong; Acharya, Lipi; Du, Lisa; Vogt, Peter K; Ping, Peipei; Yates, John R

    2015-11-03

    Shotgun proteomics generates valuable information from large-scale and target protein characterizations, including protein expression, protein quantification, protein post-translational modifications (PTMs), protein localization, and protein-protein interactions. Typically, peptides derived from proteolytic digestion, rather than intact proteins, are analyzed by mass spectrometers because peptides are more readily separated, ionized and fragmented. The amino acid sequences of peptides can be interpreted by matching the observed tandem mass spectra to theoretical spectra derived from a protein sequence database. Identified peptides serve as surrogates for their proteins and are often used to establish what proteins were present in the original mixture and to quantify protein abundance. Two major issues exist for assigning peptides to their originating protein. The first issue is maintaining a desired false discovery rate (FDR) when comparing or combining multiple large datasets generated by shotgun analysis and the second issue is properly assigning peptides to proteins when homologous proteins are present in the database. Herein we demonstrate a new computational tool, ProteinInferencer, which can be used for protein inference with both small- or large-scale data sets to produce a well-controlled protein FDR. In addition, ProteinInferencer introduces confidence scoring for individual proteins, which makes protein identifications evaluable. This article is part of a Special Issue entitled: Computational Proteomics. Copyright © 2015. Published by Elsevier B.V.

  3. A plea for a global natural history collection - online

    USDA-ARS?s Scientific Manuscript database

    Species are the currency of comparative biology: scientists from many biological disciplines, including community ecology, conservation biology, pest management, and biological control rely on scientifically sound, objective species data. However, large-scale species identifications are often not fe...

  4. MIPHENO: Data normalization for high throughput metabolic analysis.

    EPA Science Inventory

    High throughput methodologies such as microarrays, mass spectrometry and plate-based small molecule screens are increasingly used to facilitate discoveries from gene function to drug candidate identification. These large-scale experiments are typically carried out over the course...

  5. Discovering Coseismic Traveling Ionospheric Disturbances Generated by the 2016 Kaikoura Earthquake

    NASA Astrophysics Data System (ADS)

    Li, J. D.; Rude, C. M.; Gowanlock, M.; Pankratius, V.

    2017-12-01

    Geophysical events and hazards, such as earthquakes, tsunamis, and volcanoes, have been shown to generate traveling ionospheric disturbances (TIDs). These disturbances can be measured by means of Total Electron Content fluctuations obtained from a network of multifrequency GPS receivers in the MIT Haystack Observatory Madrigal database. Analyzing the response of the ionosphere to such hazards enhances our understanding of natural phenomena and augments our large-scale monitoring capabilities in conjunction with other ground-based sensors. However, it is currently challenging for human investigators to spot and characterize such signatures, or whether a geophysical event has actually occurred, because the ionosphere can be noisy with multiple simultaneous phenomena taking place at the same time. This work therefore explores a systematic pipeline for the ex-post discovery and characterization of TIDs. Our technique starts by geolocating the event and gathering the corresponding data, then checks for potentially conflicting TID sources, and processes the raw total electron content data to generate differential measurements. A Kolmogorov-Smirnov test is applied to evaluate the statistical significance of detected deviations in the differential measurements. We present results from our successful application of this pipeline to the 2016 7.8 Mw Kaikoura earthquake occurring in New Zealand on November 13th. We detect a coseismic TID occurring 8 minutes after the earthquake and propagating towards the equator at 1050 m/s, with a 0.22 peak-to-peak TECu amplitude. Furthermore, the observed waveform exhibits more complex behavior than the expected N-wave for a coseismic TID, which potentially results from the complex multi-fault structure of the earthquake. We acknowledge support from NSF ACI1442997 (PI Pankratius), NASA AISTNNX15AG84G (PI Pankratius), and NSF AGS-1343967 (PI Pankratius), and NSF AGS-1242204 (PI Erickson).

  6. New Az/El mount for Haystack Observatory's Small Radio Telescope kit

    NASA Astrophysics Data System (ADS)

    Cobb, M. L.

    2005-12-01

    The Small Radio Telescope (SRT) kit was designed by Haystack Observatory as part of their educational outreach effort. The SRT uses a custom designed FFT based radio spectrometer receiver with a controller to position a 2.3m dish to make various radio astronomy observations including the 21 cm spin flip line of atomic hydrogen. Because there is no sizable commercial market for a two dimensional mount for dishes of this size, finding an appropriate provider as been a recurring problem for the project. Originally, the kit used a modified motor mount from Kaultronics called the H180. Two of these motors were combined by a specially designed adaptor to allow motion in azimuth and elevation. When Kaultronics was bought out by California Amplifier they discontinued production of the H180. The next iteration used a compact unit called the alfa-spid which was made in Germany and imported through Canada. The alfa-spid was designed to point various ham radio antennas and proved problematic with 2.3m dishes. Most recently the CASSI (Custom Astronomical Support Services, Inc.) corporation has designed and certified a robust Az/El mount capable of supporting dishes up to 12 feet (3.6m) with 100 MPH wind loads. This paper presents the design and operating characteristics of the new CASSI mount. The CASSI mount is now shipped with the SRT kit and should serve the project well for the foreseeable future.

  7. Automatic Selection of Order Parameters in the Analysis of Large Scale Molecular Dynamics Simulations.

    PubMed

    Sultan, Mohammad M; Kiss, Gert; Shukla, Diwakar; Pande, Vijay S

    2014-12-09

    Given the large number of crystal structures and NMR ensembles that have been solved to date, classical molecular dynamics (MD) simulations have become powerful tools in the atomistic study of the kinetics and thermodynamics of biomolecular systems on ever increasing time scales. By virtue of the high-dimensional conformational state space that is explored, the interpretation of large-scale simulations faces difficulties not unlike those in the big data community. We address this challenge by introducing a method called clustering based feature selection (CB-FS) that employs a posterior analysis approach. It combines supervised machine learning (SML) and feature selection with Markov state models to automatically identify the relevant degrees of freedom that separate conformational states. We highlight the utility of the method in the evaluation of large-scale simulations and show that it can be used for the rapid and automated identification of relevant order parameters involved in the functional transitions of two exemplary cell-signaling proteins central to human disease states.

  8. The consequences of landscape change on ecological resources: An assessment of the United States Mid-Atlantic region, 1973-1993

    Treesearch

    K. Bruce Jones; Anne C. Neale; Timothy G. Wade; James D. Wickham; Chad L. Cross; Curtis M. Edmonds; Thomas R. Loveland; Maliha S. Nash; Kurt H. Riitters; Elizabeth R. Smith

    2001-01-01

    Spatially explicit identification of changes in ecological conditions over large areas is key to targeting and prioitizing areas for environmental protection and restoration by managers at watershed, basin, and regional scales. A critical limitation to this point has been the development of methods to conduct such broad-scale assessments. Field-based methods have...

  9. Research on large-scale wind farm modeling

    NASA Astrophysics Data System (ADS)

    Ma, Longfei; Zhang, Baoqun; Gong, Cheng; Jiao, Ran; Shi, Rui; Chi, Zhongjun; Ding, Yifeng

    2017-01-01

    Due to intermittent and adulatory properties of wind energy, when large-scale wind farm connected to the grid, it will have much impact on the power system, which is different from traditional power plants. Therefore it is necessary to establish an effective wind farm model to simulate and analyze the influence wind farms have on the grid as well as the transient characteristics of the wind turbines when the grid is at fault. However we must first establish an effective WTGs model. As the doubly-fed VSCF wind turbine has become the mainstream wind turbine model currently, this article first investigates the research progress of doubly-fed VSCF wind turbine, and then describes the detailed building process of the model. After that investigating the common wind farm modeling methods and pointing out the problems encountered. As WAMS is widely used in the power system, which makes online parameter identification of the wind farm model based on off-output characteristics of wind farm be possible, with a focus on interpretation of the new idea of identification-based modeling of large wind farms, which can be realized by two concrete methods.

  10. The PREP pipeline: standardized preprocessing for large-scale EEG analysis.

    PubMed

    Bigdely-Shamlo, Nima; Mullen, Tim; Kothe, Christian; Su, Kyung-Min; Robbins, Kay A

    2015-01-01

    The technology to collect brain imaging and physiological measures has become portable and ubiquitous, opening the possibility of large-scale analysis of real-world human imaging. By its nature, such data is large and complex, making automated processing essential. This paper shows how lack of attention to the very early stages of an EEG preprocessing pipeline can reduce the signal-to-noise ratio and introduce unwanted artifacts into the data, particularly for computations done in single precision. We demonstrate that ordinary average referencing improves the signal-to-noise ratio, but that noisy channels can contaminate the results. We also show that identification of noisy channels depends on the reference and examine the complex interaction of filtering, noisy channel identification, and referencing. We introduce a multi-stage robust referencing scheme to deal with the noisy channel-reference interaction. We propose a standardized early-stage EEG processing pipeline (PREP) and discuss the application of the pipeline to more than 600 EEG datasets. The pipeline includes an automatically generated report for each dataset processed. Users can download the PREP pipeline as a freely available MATLAB library from http://eegstudy.org/prepcode.

  11. A cosmic superfluid phase

    NASA Technical Reports Server (NTRS)

    Gradwohl, Ben-Ami

    1991-01-01

    The universe may have undergone a superfluid-like phase during its evolution, resulting from the injection of nontopological charge into the spontaneously broken vacuum. In the presence of vortices this charge is identified with angular momentum. This leads to turbulent domains on the scale of the correlation length. By restoring the symmetry at low temperatures, the vortices dissociate and push the charges to the boundaries of these domains. The model can be scaled (phenomenologically) to very low energies, it can be incorporated in a late time phase transition and form large scale structure in the boundary layers of the correlation volumes. The novel feature of the model lies in the fact that the dark matter is endowed with coherent motion. The possibilities of identifying this flow around superfluid vortices with the observed large scale bulk motion is discussed. If this identification is possible, then the definite prediction can be made that a more extended map of peculiar velocities would have to reveal large scale circulations in the flow pattern.

  12. Computer aided manual validation of mass spectrometry-based proteomic data.

    PubMed

    Curran, Timothy G; Bryson, Bryan D; Reigelhaupt, Michael; Johnson, Hannah; White, Forest M

    2013-06-15

    Advances in mass spectrometry-based proteomic technologies have increased the speed of analysis and the depth provided by a single analysis. Computational tools to evaluate the accuracy of peptide identifications from these high-throughput analyses have not kept pace with technological advances; currently the most common quality evaluation methods are based on statistical analysis of the likelihood of false positive identifications in large-scale data sets. While helpful, these calculations do not consider the accuracy of each identification, thus creating a precarious situation for biologists relying on the data to inform experimental design. Manual validation is the gold standard approach to confirm accuracy of database identifications, but is extremely time-intensive. To palliate the increasing time required to manually validate large proteomic datasets, we provide computer aided manual validation software (CAMV) to expedite the process. Relevant spectra are collected, catalogued, and pre-labeled, allowing users to efficiently judge the quality of each identification and summarize applicable quantitative information. CAMV significantly reduces the burden associated with manual validation and will hopefully encourage broader adoption of manual validation in mass spectrometry-based proteomics. Copyright © 2013 Elsevier Inc. All rights reserved.

  13. Banana production systems: identification of alternative systems for more sustainable production.

    PubMed

    Bellamy, Angelina Sanderson

    2013-04-01

    Large-scale, monoculture production systems dependent on synthetic fertilizers and pesticides, increase yields, but are costly and have deleterious impacts on human health and the environment. This research investigates variations in banana production practices in Costa Rica, to identify alternative systems that combine high productivity and profitability, with reduced reliance on agrochemicals. Farm workers were observed during daily production activities; 39 banana producers and 8 extension workers/researchers were interviewed; and a review of field experiments conducted by the National Banana Corporation between 1997 and 2002 was made. Correspondence analysis showed that there is no structured variation in large-scale banana producers' practices, but two other banana production systems were identified: a small-scale organic system and a small-scale conventional coffee-banana intercropped system. Field-scale research may reveal ways that these practices can be scaled up to achieve a productive and profitable system producing high-quality export bananas with fewer or no pesticides.

  14. A new way to protect privacy in large-scale genome-wide association studies.

    PubMed

    Kamm, Liina; Bogdanov, Dan; Laur, Sven; Vilo, Jaak

    2013-04-01

    Increased availability of various genotyping techniques has initiated a race for finding genetic markers that can be used in diagnostics and personalized medicine. Although many genetic risk factors are known, key causes of common diseases with complex heritage patterns are still unknown. Identification of such complex traits requires a targeted study over a large collection of data. Ideally, such studies bring together data from many biobanks. However, data aggregation on such a large scale raises many privacy issues. We show how to conduct such studies without violating privacy of individual donors and without leaking the data to third parties. The presented solution has provable security guarantees. Supplementary data are available at Bioinformatics online.

  15. Finding needles in haystacks: linking scientific names, reference specimens and molecular data for Fungi.

    PubMed

    Schoch, Conrad L; Robbertse, Barbara; Robert, Vincent; Vu, Duong; Cardinali, Gianluigi; Irinyi, Laszlo; Meyer, Wieland; Nilsson, R Henrik; Hughes, Karen; Miller, Andrew N; Kirk, Paul M; Abarenkov, Kessy; Aime, M Catherine; Ariyawansa, Hiran A; Bidartondo, Martin; Boekhout, Teun; Buyck, Bart; Cai, Qing; Chen, Jie; Crespo, Ana; Crous, Pedro W; Damm, Ulrike; De Beer, Z Wilhelm; Dentinger, Bryn T M; Divakar, Pradeep K; Dueñas, Margarita; Feau, Nicolas; Fliegerova, Katerina; García, Miguel A; Ge, Zai-Wei; Griffith, Gareth W; Groenewald, Johannes Z; Groenewald, Marizeth; Grube, Martin; Gryzenhout, Marieka; Gueidan, Cécile; Guo, Liangdong; Hambleton, Sarah; Hamelin, Richard; Hansen, Karen; Hofstetter, Valérie; Hong, Seung-Beom; Houbraken, Jos; Hyde, Kevin D; Inderbitzin, Patrik; Johnston, Peter R; Karunarathna, Samantha C; Kõljalg, Urmas; Kovács, Gábor M; Kraichak, Ekaphan; Krizsan, Krisztina; Kurtzman, Cletus P; Larsson, Karl-Henrik; Leavitt, Steven; Letcher, Peter M; Liimatainen, Kare; Liu, Jian-Kui; Lodge, D Jean; Luangsa-ard, Janet Jennifer; Lumbsch, H Thorsten; Maharachchikumbura, Sajeewa S N; Manamgoda, Dimuthu; Martín, María P; Minnis, Andrew M; Moncalvo, Jean-Marc; Mulè, Giuseppina; Nakasone, Karen K; Niskanen, Tuula; Olariaga, Ibai; Papp, Tamás; Petkovits, Tamás; Pino-Bodas, Raquel; Powell, Martha J; Raja, Huzefa A; Redecker, Dirk; Sarmiento-Ramirez, J M; Seifert, Keith A; Shrestha, Bhushan; Stenroos, Soili; Stielow, Benjamin; Suh, Sung-Oui; Tanaka, Kazuaki; Tedersoo, Leho; Telleria, M Teresa; Udayanga, Dhanushka; Untereiner, Wendy A; Diéguez Uribeondo, Javier; Subbarao, Krishna V; Vágvölgyi, Csaba; Visagie, Cobus; Voigt, Kerstin; Walker, Donald M; Weir, Bevan S; Weiß, Michael; Wijayawardene, Nalin N; Wingfield, Michael J; Xu, J P; Yang, Zhu L; Zhang, Ning; Zhuang, Wen-Ying; Federhen, Scott

    2014-01-01

    DNA phylogenetic comparisons have shown that morphology-based species recognition often underestimates fungal diversity. Therefore, the need for accurate DNA sequence data, tied to both correct taxonomic names and clearly annotated specimen data, has never been greater. Furthermore, the growing number of molecular ecology and microbiome projects using high-throughput sequencing require fast and effective methods for en masse species assignments. In this article, we focus on selecting and re-annotating a set of marker reference sequences that represent each currently accepted order of Fungi. The particular focus is on sequences from the internal transcribed spacer region in the nuclear ribosomal cistron, derived from type specimens and/or ex-type cultures. Re-annotated and verified sequences were deposited in a curated public database at the National Center for Biotechnology Information (NCBI), namely the RefSeq Targeted Loci (RTL) database, and will be visible during routine sequence similarity searches with NR_prefixed accession numbers. A set of standards and protocols is proposed to improve the data quality of new sequences, and we suggest how type and other reference sequences can be used to improve identification of Fungi. Database URL: http://www.ncbi.nlm.nih.gov/bioproject/PRJNA177353. Published by Oxford University Press 2013. This work is written by US Government employees and is in the public domain in the US.

  16. Identification of differentially methylated sites with weak methylation effect

    USDA-ARS?s Scientific Manuscript database

    DNA methylation is an epigenetic alteration crucial for regulating stress responses. Identifying large-scale DNA methylation at single nucleotide resolution is made possible by whole genome bisulfite sequencing. An essential task following the generation of bisulfite sequencing data is to detect dif...

  17. Torching the Haystack: modelling fast-fail strategies in drug development.

    PubMed

    Lendrem, Dennis W; Lendrem, B Clare

    2013-04-01

    By quickly clearing the development pipeline of failing or marginal products, fast-fail strategies release resources to focus on more promising molecules. The Quick-Kill model of drug development demonstrates that fast-fail strategies will: (1) reduce the expected time to market; (2) reduce expected R&D costs; and (3) increase R&D productivity. This paper outlines the model and demonstrates the impact of fast-fail strategies. The model is illustrated with costs and risks data from pharmaceutical and biopharmaceutical companies. Copyright © 2012 Elsevier Ltd. All rights reserved.

  18. Robust Computation of Linear Models, or How to Find a Needle in a Haystack

    DTIC Science & Technology

    2012-02-17

    robustly, project it onto a sphere, and then apply standard PCA. This approach is due to [LMS+99]. Maronna et al . [MMY06] recommend it as a preferred...of this form is due to Chandrasekaran et al . [CSPW11]. Given an observed matrix X, they propose to solve the semidefinite problem minimize ‖P ‖S1 + γ...regularization parameter γ negotiates a tradeoff between the two goals. Candès et al . [CLMW11] study the performance of (2.1) for robust linear

  19. Iris indexing based on local intensity order pattern

    NASA Astrophysics Data System (ADS)

    Emerich, Simina; Malutan, Raul; Crisan, Septimiu; Lefkovits, Laszlo

    2017-03-01

    In recent years, iris biometric systems have increased in popularity and have been proven that are capable of handling large-scale databases. The main advantage of these systems is accuracy and reliability. A proper iris patterns classification is expected to reduce the matching time in huge databases. This paper presents an iris indexing technique based on Local Intensity Order Pattern. The performance of the present approach is evaluated on UPOL database and is compared with other recent systems designed for iris indexing. The results illustrate the potential of the proposed method for large scale iris identification.

  20. Fast Open-World Person Re-Identification.

    PubMed

    Zhu, Xiatian; Wu, Botong; Huang, Dongcheng; Zheng, Wei-Shi

    2018-05-01

    Existing person re-identification (re-id) methods typically assume that: 1) any probe person is guaranteed to appear in the gallery target population during deployment (i.e., closed-world) and 2) the probe set contains only a limited number of people (i.e., small search scale). Both assumptions are artificial and breached in real-world applications, since the probe population in target people search can be extremely vast in practice due to the ambiguity of probe search space boundary. Therefore, it is unrealistic that any probe person is assumed as one target people, and a large-scale search in person images is inherently demanded. In this paper, we introduce a new person re-id search setting, called large scale open-world (LSOW) re-id, characterized by huge size probe images and open person population in search thus more close to practical deployments. Under LSOW, the under-studied problem of person re-id efficiency is essential in addition to that of commonly studied re-id accuracy. We, therefore, develop a novel fast person re-id method, called Cross-view Identity Correlation and vErification (X-ICE) hashing, for joint learning of cross-view identity representation binarisation and discrimination in a unified manner. Extensive comparative experiments on three large-scale benchmarks have been conducted to validate the superiority and advantages of the proposed X-ICE method over a wide range of the state-of-the-art hashing models, person re-id methods, and their combinations.

  1. A Needle in a Haystack? Uninsured Workers in Small Businesses That Do Not Offer Coverage

    PubMed Central

    Kronick, Richard; Olsen, Louis C

    2006-01-01

    Objective To describe the insurance status of workers at small businesses, and to describe the status of uninsured persons by the employment characteristics (employment status, firm size, and whether the employer offers insurance) of the head of household. Data Sources Data from the March and February 2001 Current Population Survey, and a survey of 2,830 small businesses in San Diego County conducted in 2001. Study Design The survey of small businesses was undertaken as part of a project testing the response of employers to offers of subsidized coverage. Employers were asked whether they offered insurance, and about the insurance status of their employees. The merged February–March 2001 CPS was used to identify the employment status, firm size, and employer-offering status for uninsured persons in the U.S. Data Collection Telephone interviews with small businesses in San Diego County. Principal Findings Only 21 percent of the uninsured in the U.S. are full-time employees (or their dependents) in small businesses (<100 employees) that do not offer insurance. The employment status of the uninsured is heterogeneous: many work for large employers, small employers who do offer insurance, or are self-employed, part-time workers, or have no workers in the household. Although there are many small businesses in San Diego that do not offer coverage, most of them have very few uninsured workers. Over 50 percent of businesses that do not offer coverage have either zero or one uninsured worker. There are very few small businesses that do not offer coverage and that have substantial numbers of uninsured workers. These businesses are not quite as rare as a needle in a haystack, but they are very difficult to find. Conclusions If all small businesses that do not offer insurance now could be persuaded to start offering coverage, and if all the full-time workers (and their dependents) in those businesses accepted insurance, the number of uninsured would decline by 21 percent—a significant decline, but leaving 80 percent of the problem untouched. If the prime target for programs of subsidized insurance are small businesses that do not offer coverage now and that have substantial numbers of uninsured workers, the target is very small. PMID:16430600

  2. Modeling of LEO Orbital Debris Populations in Centimeter and Millimeter Size Regimes

    NASA Technical Reports Server (NTRS)

    Xu, Y.-L.; Hill, . M.; Horstman, M.; Krisko, P. H.; Liou, J.-C.; Matney, M.; Stansbery, E. G.

    2010-01-01

    The building of the NASA Orbital Debris Engineering Model, whether ORDEM2000 or its recently updated version ORDEM2010, uses as its foundation a number of model debris populations, each truncated at a minimum object-size ranging from 10 micron to 1 m. This paper discusses the development of the ORDEM2010 model debris populations in LEO (low Earth orbit), focusing on centimeter (smaller than 10 cm) and millimeter size regimes. Primary data sets used in the statistical derivation of the cm- and mm-size model populations are from the Haystack radar operated in a staring mode. Unlike cataloged objects of sizes greater than approximately 10 cm, ground-based radars monitor smaller-size debris only in a statistical manner instead of tracking every piece. The mono-static Haystack radar can detect debris as small as approximately 5 mm at moderate LEO altitudes. Estimation of millimeter debris populations (for objects smaller than approximately 6 mm) rests largely on Goldstone radar measurements. The bi-static Goldstone radar can detect 2- to 3-mm objects. The modeling of the cm- and mm-debris populations follows the general approach to developing other ORDEM2010-required model populations for various components and types of debris. It relies on appropriate reference populations to provide necessary prior information on the orbital structures and other important characteristics of the debris objects. NASA's LEO-to-GEO Environment Debris (LEGEND) model is capable of furnishing such reference populations in the desired size range. A Bayesian statistical inference process, commonly adopted in ORDEM2010 model-population derivations, changes a priori distribution into a posteriori distribution and thus refines the reference populations in terms of data. This paper describes key elements and major steps in the statistical derivations of the cm- and mm-size debris populations and presents results. Due to lack of data for near 1-mm sizes, the model populations of 1- to 3.16-mm objects are an empirical extension from larger debris. The extension takes into account the results of micro-debris (from 10 micron to 1 mm) population modeling that is based on shuttle impact data, in the hope of making a smooth transition between micron and millimeter size regimes. This paper also includes a brief discussion on issues and potential future work concerning the analysis and interpretation of Goldstone radar data.

  3. Large-Scale Off-Target Identification Using Fast and Accurate Dual Regularized One-Class Collaborative Filtering and Its Application to Drug Repurposing.

    PubMed

    Lim, Hansaim; Poleksic, Aleksandar; Yao, Yuan; Tong, Hanghang; He, Di; Zhuang, Luke; Meng, Patrick; Xie, Lei

    2016-10-01

    Target-based screening is one of the major approaches in drug discovery. Besides the intended target, unexpected drug off-target interactions often occur, and many of them have not been recognized and characterized. The off-target interactions can be responsible for either therapeutic or side effects. Thus, identifying the genome-wide off-targets of lead compounds or existing drugs will be critical for designing effective and safe drugs, and providing new opportunities for drug repurposing. Although many computational methods have been developed to predict drug-target interactions, they are either less accurate than the one that we are proposing here or computationally too intensive, thereby limiting their capability for large-scale off-target identification. In addition, the performances of most machine learning based algorithms have been mainly evaluated to predict off-target interactions in the same gene family for hundreds of chemicals. It is not clear how these algorithms perform in terms of detecting off-targets across gene families on a proteome scale. Here, we are presenting a fast and accurate off-target prediction method, REMAP, which is based on a dual regularized one-class collaborative filtering algorithm, to explore continuous chemical space, protein space, and their interactome on a large scale. When tested in a reliable, extensive, and cross-gene family benchmark, REMAP outperforms the state-of-the-art methods. Furthermore, REMAP is highly scalable. It can screen a dataset of 200 thousands chemicals against 20 thousands proteins within 2 hours. Using the reconstructed genome-wide target profile as the fingerprint of a chemical compound, we predicted that seven FDA-approved drugs can be repurposed as novel anti-cancer therapies. The anti-cancer activity of six of them is supported by experimental evidences. Thus, REMAP is a valuable addition to the existing in silico toolbox for drug target identification, drug repurposing, phenotypic screening, and side effect prediction. The software and benchmark are available at https://github.com/hansaimlim/REMAP.

  4. Large-Scale Off-Target Identification Using Fast and Accurate Dual Regularized One-Class Collaborative Filtering and Its Application to Drug Repurposing

    PubMed Central

    Poleksic, Aleksandar; Yao, Yuan; Tong, Hanghang; Meng, Patrick; Xie, Lei

    2016-01-01

    Target-based screening is one of the major approaches in drug discovery. Besides the intended target, unexpected drug off-target interactions often occur, and many of them have not been recognized and characterized. The off-target interactions can be responsible for either therapeutic or side effects. Thus, identifying the genome-wide off-targets of lead compounds or existing drugs will be critical for designing effective and safe drugs, and providing new opportunities for drug repurposing. Although many computational methods have been developed to predict drug-target interactions, they are either less accurate than the one that we are proposing here or computationally too intensive, thereby limiting their capability for large-scale off-target identification. In addition, the performances of most machine learning based algorithms have been mainly evaluated to predict off-target interactions in the same gene family for hundreds of chemicals. It is not clear how these algorithms perform in terms of detecting off-targets across gene families on a proteome scale. Here, we are presenting a fast and accurate off-target prediction method, REMAP, which is based on a dual regularized one-class collaborative filtering algorithm, to explore continuous chemical space, protein space, and their interactome on a large scale. When tested in a reliable, extensive, and cross-gene family benchmark, REMAP outperforms the state-of-the-art methods. Furthermore, REMAP is highly scalable. It can screen a dataset of 200 thousands chemicals against 20 thousands proteins within 2 hours. Using the reconstructed genome-wide target profile as the fingerprint of a chemical compound, we predicted that seven FDA-approved drugs can be repurposed as novel anti-cancer therapies. The anti-cancer activity of six of them is supported by experimental evidences. Thus, REMAP is a valuable addition to the existing in silico toolbox for drug target identification, drug repurposing, phenotypic screening, and side effect prediction. The software and benchmark are available at https://github.com/hansaimlim/REMAP. PMID:27716836

  5. A Decision Mixture Model-Based Method for Inshore Ship Detection Using High-Resolution Remote Sensing Images

    PubMed Central

    Bi, Fukun; Chen, Jing; Zhuang, Yin; Bian, Mingming; Zhang, Qingjun

    2017-01-01

    With the rapid development of optical remote sensing satellites, ship detection and identification based on large-scale remote sensing images has become a significant maritime research topic. Compared with traditional ocean-going vessel detection, inshore ship detection has received increasing attention in harbor dynamic surveillance and maritime management. However, because the harbor environment is complex, gray information and texture features between docked ships and their connected dock regions are indistinguishable, most of the popular detection methods are limited by their calculation efficiency and detection accuracy. In this paper, a novel hierarchical method that combines an efficient candidate scanning strategy and an accurate candidate identification mixture model is presented for inshore ship detection in complex harbor areas. First, in the candidate region extraction phase, an omnidirectional intersected two-dimension scanning (OITDS) strategy is designed to rapidly extract candidate regions from the land-water segmented images. In the candidate region identification phase, a decision mixture model (DMM) is proposed to identify real ships from candidate objects. Specifically, to improve the robustness regarding the diversity of ships, a deformable part model (DPM) was employed to train a key part sub-model and a whole ship sub-model. Furthermore, to improve the identification accuracy, a surrounding correlation context sub-model is built. Finally, to increase the accuracy of candidate region identification, these three sub-models are integrated into the proposed DMM. Experiments were performed on numerous large-scale harbor remote sensing images, and the results showed that the proposed method has high detection accuracy and rapid computational efficiency. PMID:28640236

  6. A Decision Mixture Model-Based Method for Inshore Ship Detection Using High-Resolution Remote Sensing Images.

    PubMed

    Bi, Fukun; Chen, Jing; Zhuang, Yin; Bian, Mingming; Zhang, Qingjun

    2017-06-22

    With the rapid development of optical remote sensing satellites, ship detection and identification based on large-scale remote sensing images has become a significant maritime research topic. Compared with traditional ocean-going vessel detection, inshore ship detection has received increasing attention in harbor dynamic surveillance and maritime management. However, because the harbor environment is complex, gray information and texture features between docked ships and their connected dock regions are indistinguishable, most of the popular detection methods are limited by their calculation efficiency and detection accuracy. In this paper, a novel hierarchical method that combines an efficient candidate scanning strategy and an accurate candidate identification mixture model is presented for inshore ship detection in complex harbor areas. First, in the candidate region extraction phase, an omnidirectional intersected two-dimension scanning (OITDS) strategy is designed to rapidly extract candidate regions from the land-water segmented images. In the candidate region identification phase, a decision mixture model (DMM) is proposed to identify real ships from candidate objects. Specifically, to improve the robustness regarding the diversity of ships, a deformable part model (DPM) was employed to train a key part sub-model and a whole ship sub-model. Furthermore, to improve the identification accuracy, a surrounding correlation context sub-model is built. Finally, to increase the accuracy of candidate region identification, these three sub-models are integrated into the proposed DMM. Experiments were performed on numerous large-scale harbor remote sensing images, and the results showed that the proposed method has high detection accuracy and rapid computational efficiency.

  7. Identification and Screening of Carcass Pretreatment ...

    EPA Pesticide Factsheets

    Technical Fact Sheet Managing the treatment and disposal of large numbers of animal carcasses following a foreign animal disease (FAD) outbreak is a challenging endeavor. Pretreatment of the infectious carcasses might facilitate the disposal of the carcasses by simplifying the transportation, reducing the pathogen load in the carcasses, or by isolating the pathogen from the environment to minimize spread of any pathogens.This brief summarizes information contained in U.S. Environmental Protection Agency (EPA) report (EPA/600/R-15/053) entitled Identification and Screening of Infectious Carcass Pretreatment Alternatives. This brief describes how each of eleven pretreatment methods can be used prior to, and in conjunction with, six commonly used large-scale carcass disposal options

  8. Chem Ed Compacts.

    ERIC Educational Resources Information Center

    Wolf, Walter A., Ed.

    1978-01-01

    Reported here are brief descriptions of a common grading and scaling formula for large multi-section courses, an ion exchange amino acid separation and thin layer chromatography identification experiment, a conservation of energy demonstration, a catalyst for synthesizing esters from fatty aids, and an inexpensive method for preparing platinum…

  9. Identification of the key ecological factors influencing vegetation degradation in semi-arid agro-pastoral ecotone considering spatial scales

    NASA Astrophysics Data System (ADS)

    Peng, Yu; Wang, Qinghui; Fan, Min

    2017-11-01

    When assessing re-vegetation project performance and optimizing land management, identification of the key ecological factors inducing vegetation degradation has crucial implications. Rainfall, temperature, elevation, slope, aspect, land use type, and human disturbance are ecological factors affecting the status of vegetation index. However, at different spatial scales, the key factors may vary. Using Helin County, Inner-Mongolia, China as the study site and combining remote sensing image interpretation, field surveying, and mathematical methods, this study assesses key ecological factors affecting vegetation degradation under different spatial scales in a semi-arid agro-pastoral ecotone. It indicates that the key factors are different at various spatial scales. Elevation, rainfall, and temperature are identified as crucial for all spatial extents. Elevation, rainfall and human disturbance are key factors for small-scale quadrats of 300 m × 300 m and 600 m × 600 m, temperature and land use type are key factors for a medium-scale quadrat of 1 km × 1 km, and rainfall, temperature, and land use are key factors for large-scale quadrats of 2 km × 2 km and 5 km × 5 km. For this region, human disturbance is not the key factor for vegetation degradation across spatial scales. It is necessary to consider spatial scale for the identification of key factors determining vegetation characteristics. The eco-restoration programs at various spatial scales should identify key influencing factors according their scales so as to take effective measurements. The new understanding obtained in this study may help to explore the forces which driving vegetation degradation in the degraded regions in the world.

  10. Identification of MAPK Substrates Using Quantitative Phosphoproteomics.

    PubMed

    Zhang, Tong; Schneider, Jacqueline D; Zhu, Ning; Chen, Sixue

    2017-01-01

    Activation of mitogen-activated protein kinases (MAPKs) under diverse biotic and abiotic factors and identification of an array of downstream MAPK target proteins are hot topics in plant signal transduction. Through interactions with a plethora of substrate proteins, MAPK cascades regulate many physiological processes in the course of plant growth, development, and response to environmental factors. Identification and quantification of potential MAPK substrates are essential, but have been technically challenging. With the recent advancement in phosphoproteomics, here we describe a method that couples metal dioxide for phosphopeptide enrichment with tandem mass tags (TMT) mass spectrometry (MS) for large-scale MAPK substrate identification and quantification. We have applied this method to a transient expression system carrying a wild type (WT) and a constitutive active (CA) version of a MAPK. This combination of genetically engineered MAPKs and phosphoproteomics provides a high-throughput, unbiased analysis of MAPK-triggered phosphorylation changes on the proteome scale. Therefore, it is a robust method for identifying potential MAPK substrates and should be applicable in the study of other kinase cascades in plants as well as in other organisms.

  11. Online video game addiction: identification of addicted adolescent gamers.

    PubMed

    Van Rooij, Antonius J; Schoenmakers, Tim M; Vermulst, Ad A; Van den Eijnden, Regina J J M; Van de Mheen, Dike

    2011-01-01

    To provide empirical data-driven identification of a group of addicted online gamers. Repeated cross-sectional survey study, comprising a longitudinal cohort, conducted in 2008 and 2009. Secondary schools in the Netherlands.   Two large samples of Dutch schoolchildren (aged 13-16 years). Compulsive internet use scale, weekly hours of online gaming and psychosocial variables. This study confirms the existence of a small group of addicted online gamers (3%), representing about 1.5% of all children aged 13-16 years in the Netherlands. Although these gamers report addiction-like problems, relationships with decreased psychosocial health were less evident. The identification of a small group of addicted online gamers supports efforts to develop and validate questionnaire scales aimed at measuring the phenomenon of online video game addiction. The findings contribute to the discussion on the inclusion of non-substance addictions in the proposed unified concept of 'Addiction and Related Disorders' for the DSM-V by providing indirect identification and validation of a group of suspected online video game addicts. © 2010 The Authors, Addiction © 2010 Society for the Study of Addiction.

  12. Bio-inspired digital signal processing for fast radionuclide mixture identification

    NASA Astrophysics Data System (ADS)

    Thevenin, M.; Bichler, O.; Thiam, C.; Bobin, C.; Lourenço, V.

    2015-05-01

    Countries are trying to equip their public transportation infrastructure with fixed radiation portals and detectors to detect radiological threat. Current works usually focus on neutron detection, which could be useless in the case of dirty bomb that would not use fissile material. Another approach, such as gamma dose rate variation monitoring is a good indication of the presence of radionuclide. However, some legitimate products emit large quantities of natural gamma rays; environment also emits gamma rays naturally. They can lead to false detections. Moreover, such radio-activity could be used to hide a threat such as material to make a dirty bomb. Consequently, radionuclide identification is a requirement and is traditionally performed by gamma spectrometry using unique spectral signature of each radionuclide. These approaches require high-resolution detectors, sufficient integration time to get enough statistics and large computing capacities for data analysis. High-resolution detectors are fragile and costly, making them bad candidates for large scale homeland security applications. Plastic scintillator and NaI detectors fit with such applications but their resolution makes identification difficult, especially radionuclides mixes. This paper proposes an original signal processing strategy based on artificial spiking neural networks to enable fast radionuclide identification at low count rate and for mixture. It presents results obtained for different challenging mixtures of radionuclides using a NaI scintillator. Results show that a correct identification is performed with less than hundred counts and no false identification is reported, enabling quick identification of a moving threat in a public transportation. Further work will focus on using plastic scintillators.

  13. Molecular Identification of XY Sex-Reversed Female and YY Male Channel Catfish

    USDA-ARS?s Scientific Manuscript database

    Production of channel catfish leads U.S. aquaculture, and monosex culture may provide higher production efficiencies. Determination of phenotypic sex is labor intensive and not practical for large scale culture. Catfish have an X-Y sex determination system with monomorphic sex chromosomes. Hormonal...

  14. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rodrigues, Davi C.; Piattella, Oliver F.; Chauvineau, Bertrand, E-mail: davi.rodrigues@cosmo-ufes.org, E-mail: Bertrand.Chauvineau@oca.eu, E-mail: oliver.piattella@pq.cnpq.br

    We show that Renormalization Group extensions of the Einstein-Hilbert action for large scale physics are not, in general, a particular case of standard Scalar-Tensor (ST) gravity. We present a new class of ST actions, in which the potential is not necessarily fixed at the action level, and show that this extended ST theory formally contains the Renormalization Group case. We also propose here a Renormalization Group scale setting identification that is explicitly covariant and valid for arbitrary relativistic fluids.

  15. The PREP pipeline: standardized preprocessing for large-scale EEG analysis

    PubMed Central

    Bigdely-Shamlo, Nima; Mullen, Tim; Kothe, Christian; Su, Kyung-Min; Robbins, Kay A.

    2015-01-01

    The technology to collect brain imaging and physiological measures has become portable and ubiquitous, opening the possibility of large-scale analysis of real-world human imaging. By its nature, such data is large and complex, making automated processing essential. This paper shows how lack of attention to the very early stages of an EEG preprocessing pipeline can reduce the signal-to-noise ratio and introduce unwanted artifacts into the data, particularly for computations done in single precision. We demonstrate that ordinary average referencing improves the signal-to-noise ratio, but that noisy channels can contaminate the results. We also show that identification of noisy channels depends on the reference and examine the complex interaction of filtering, noisy channel identification, and referencing. We introduce a multi-stage robust referencing scheme to deal with the noisy channel-reference interaction. We propose a standardized early-stage EEG processing pipeline (PREP) and discuss the application of the pipeline to more than 600 EEG datasets. The pipeline includes an automatically generated report for each dataset processed. Users can download the PREP pipeline as a freely available MATLAB library from http://eegstudy.org/prepcode. PMID:26150785

  16. Role of optometry school in single day large scale school vision testing

    PubMed Central

    Anuradha, N; Ramani, Krishnakumar

    2015-01-01

    Background: School vision testing aims at identification and management of refractive errors. Large-scale school vision testing using conventional methods is time-consuming and demands a lot of chair time from the eye care professionals. A new strategy involving a school of optometry in single day large scale school vision testing is discussed. Aim: The aim was to describe a new approach of performing vision testing of school children on a large scale in a single day. Materials and Methods: A single day vision testing strategy was implemented wherein 123 members (20 teams comprising optometry students and headed by optometrists) conducted vision testing for children in 51 schools. School vision testing included basic vision screening, refraction, frame measurements, frame choice and referrals for other ocular problems. Results: A total of 12448 children were screened, among whom 420 (3.37%) were identified to have refractive errors. 28 (1.26%) children belonged to the primary, 163 to middle (9.80%), 129 (4.67%) to secondary and 100 (1.73%) to the higher secondary levels of education respectively. 265 (2.12%) children were referred for further evaluation. Conclusion: Single day large scale school vision testing can be adopted by schools of optometry to reach a higher number of children within a short span. PMID:25709271

  17. De-Identification in Learning Analytics

    ERIC Educational Resources Information Center

    Khalila, Mohammad; Ebner, Martin

    2016-01-01

    Learning analytics has reserved its position as an important field in the educational sector. However, the large-scale collection, processing, and analyzing of data has steered the wheel beyond the borders to face an abundance of ethical breaches and constraints. Revealing learners' personal information and attitudes, as well as their activities,…

  18. Novel insights into host responses and the reproductive pathophysiology of type 2 porcine reproductive and respiratory syndrome (PRRS)

    USDA-ARS?s Scientific Manuscript database

    A large-scale challenge experiment using type 2 porcine reproductive and respiratory virus (PRRSV) provided new insights into the pathophysiology of reproductive PRRS in third-trimester pregnant gilts. Deep phenotyping enabled identification of maternal and fetal factors predictive of PRRS severity ...

  19. INTEGRATING FIELD-BASED SAMPLING AND LANDSCAPE DATA FOR REGIONAL SCALE ASSESSMENTS: EXAMPLES FROM THE UNITED STATES MID-ATLANTIC REGION

    EPA Science Inventory

    Spatially explicit identification of status and changes in ecological conditions over large, regional areas is key to targeting and prioritizing areas for potential further study and environmental protection and restoration. A critical limitation to this point has been our abili...

  20. Partial Identification of Treatment Effects: Applications to Generalizability

    ERIC Educational Resources Information Center

    Chan, Wendy

    2016-01-01

    Results from large-scale evaluation studies form the foundation of evidence-based policy. The randomized experiment is often considered the gold standard among study designs because the causal impact of a treatment or intervention can be assessed without threats of confounding from external variables. Policy-makers have become increasingly…

  1. THE CONSEQUENCES OF LANDSCAPE CHANGE ON ECOLOGICAL RESOURCES: AN ASSESSMENT OF THE UNITED STATES MID-ATLANTIC REGION

    EPA Science Inventory



    Spatially explicit identification of changes in ecological conditions over large areas is key to targeting and prioritizing areas for environmental protection and restoration by managers at watershed, basin, and regional scales. A critical limitation to this point has bee...

  2. A Glance at Microsatellite Motifs from 454 Sequencing Reads of Watermelon Genomic DNA

    USDA-ARS?s Scientific Manuscript database

    A single 454 (Life Sciences Sequencing Technology) run of Charleston Gray watermelon (Citrullus lanatus var. lanatus) genomic DNA was performed and sequence data were assembled. A large scale identification of simple sequence repeat (SSR) was performed and SSR sequence data were used for the develo...

  3. SPATIAL PATTERN OF FUTURE VULNERABILITY OF STREAM EUTROPHICATION IN THE MID-ATLANTIC REGION OF THE UNITED STATES

    EPA Science Inventory

    Spatially explicit identification of changes in ecological conditions over large areas is key to targeting and prioritizing areas for environmental protection and restoration by managers at watershed, basin, and regional scales. A critical limitation to this point has been the d...

  4. Parallel Clustering Algorithm for Large-Scale Biological Data Sets

    PubMed Central

    Wang, Minchao; Zhang, Wu; Ding, Wang; Dai, Dongbo; Zhang, Huiran; Xie, Hao; Chen, Luonan; Guo, Yike; Xie, Jiang

    2014-01-01

    Backgrounds Recent explosion of biological data brings a great challenge for the traditional clustering algorithms. With increasing scale of data sets, much larger memory and longer runtime are required for the cluster identification problems. The affinity propagation algorithm outperforms many other classical clustering algorithms and is widely applied into the biological researches. However, the time and space complexity become a great bottleneck when handling the large-scale data sets. Moreover, the similarity matrix, whose constructing procedure takes long runtime, is required before running the affinity propagation algorithm, since the algorithm clusters data sets based on the similarities between data pairs. Methods Two types of parallel architectures are proposed in this paper to accelerate the similarity matrix constructing procedure and the affinity propagation algorithm. The memory-shared architecture is used to construct the similarity matrix, and the distributed system is taken for the affinity propagation algorithm, because of its large memory size and great computing capacity. An appropriate way of data partition and reduction is designed in our method, in order to minimize the global communication cost among processes. Result A speedup of 100 is gained with 128 cores. The runtime is reduced from serval hours to a few seconds, which indicates that parallel algorithm is capable of handling large-scale data sets effectively. The parallel affinity propagation also achieves a good performance when clustering large-scale gene data (microarray) and detecting families in large protein superfamilies. PMID:24705246

  5. Level-set techniques for facies identification in reservoir modeling

    NASA Astrophysics Data System (ADS)

    Iglesias, Marco A.; McLaughlin, Dennis

    2011-03-01

    In this paper we investigate the application of level-set techniques for facies identification in reservoir models. The identification of facies is a geometrical inverse ill-posed problem that we formulate in terms of shape optimization. The goal is to find a region (a geologic facies) that minimizes the misfit between predicted and measured data from an oil-water reservoir. In order to address the shape optimization problem, we present a novel application of the level-set iterative framework developed by Burger in (2002 Interfaces Free Bound. 5 301-29 2004 Inverse Problems 20 259-82) for inverse obstacle problems. The optimization is constrained by (the reservoir model) a nonlinear large-scale system of PDEs that describes the reservoir dynamics. We reformulate this reservoir model in a weak (integral) form whose shape derivative can be formally computed from standard results of shape calculus. At each iteration of the scheme, the current estimate of the shape derivative is utilized to define a velocity in the level-set equation. The proper selection of this velocity ensures that the new shape decreases the cost functional. We present results of facies identification where the velocity is computed with the gradient-based (GB) approach of Burger (2002) and the Levenberg-Marquardt (LM) technique of Burger (2004). While an adjoint formulation allows the straightforward application of the GB approach, the LM technique requires the computation of the large-scale Karush-Kuhn-Tucker system that arises at each iteration of the scheme. We efficiently solve this system by means of the representer method. We present some synthetic experiments to show and compare the capabilities and limitations of the proposed implementations of level-set techniques for the identification of geologic facies.

  6. Orbital Debris Research in the United States

    NASA Technical Reports Server (NTRS)

    Stansbery, Gene

    2009-01-01

    The presentation includes information about growth of the satellite population, the U.S. Space Surveillance Network, tracking and catalog maintenance, Haystack and HAX radar observation, Goldstone radar, the Michigan Orbital Debris Survey Telescope (MODEST), spacecraft surface examinations and sample of space shuttle impacts. GEO/LEO observations from Kwajalein Atoll, NASA s Orbital Debris Engineering Model (ORDEM2008), a LEO-to-GEO Environment Debris Model (LEGEND), Debris Assessment Software (DAS) 2.0, the NASA/JSC BUMPER-II meteoroid/debris threat assessment code, satellite reentry risk assessment, optical size and shape determination, work on more complicated fragments, and spectral studies.

  7. Radar Measurements of Small Debris from HUSIR and HAX

    NASA Technical Reports Server (NTRS)

    Hamilton J.; Blackwell, C.; McSheehy, R.; Juarez, Q.; Anz-Meador, P.

    2017-01-01

    For many years, the NASA Orbital Debris Program Office has been collecting measurements of the orbital debris environment from the Haystack Ultra-wideband Satellite Imaging Radar (HUSIR) and its auxiliary (HAX). These measurements sample the small debris population in low earth orbit (LEO). This paper will provide an overview of recent observations and highlight trends in selected debris populations. Using the NASA size estimation model, objects with a characteristic size of 1 cm and larger observed from HUSIR will be presented. Also, objects with a characteristic size of 2 cm and larger observed from HAX will be presented.

  8. Identification and Characterization of Genomic Amplifications in Ovarian Serous Carcinoma

    DTIC Science & Technology

    2009-07-01

    oncogenes, Rsf1 and Notch3, which were up-regulated in both genomic DNA and transcript levels in ovarian cancer. In a large- scale FISH analysis, Rsf1...associated with worse disease outcome, suggesting that Rsf1 could be potentially used as a prognostic marker in the future (Appendix #1). For the...over- expressed in a recurrent carcinoma. Although the follow-up study in a larger- scale sample size did not demonstrate clear amplification in NAC1

  9. Radiography with cosmic-ray and compact accelerator muons; Exploring inner-structure of large-scale objects and landforms

    PubMed Central

    NAGAMINE, Kanetada

    2016-01-01

    Cosmic-ray muons (CRM) arriving from the sky on the surface of the earth are now known to be used as radiography purposes to explore the inner-structure of large-scale objects and landforms, ranging in thickness from meter to kilometers scale, such as volcanic mountains, blast furnaces, nuclear reactors etc. At the same time, by using muons produced by compact accelerators (CAM), advanced radiography can be realized for objects with a thickness in the sub-millimeter to meter range, with additional exploration capability such as element identification and bio-chemical analysis. In the present report, principles, methods and specific research examples of CRM transmission radiography are summarized after which, principles, methods and perspective views of the future CAM radiography are described. PMID:27725469

  10. Radiography with cosmic-ray and compact accelerator muons; Exploring inner-structure of large-scale objects and landforms.

    PubMed

    Nagamine, Kanetada

    2016-01-01

    Cosmic-ray muons (CRM) arriving from the sky on the surface of the earth are now known to be used as radiography purposes to explore the inner-structure of large-scale objects and landforms, ranging in thickness from meter to kilometers scale, such as volcanic mountains, blast furnaces, nuclear reactors etc. At the same time, by using muons produced by compact accelerators (CAM), advanced radiography can be realized for objects with a thickness in the sub-millimeter to meter range, with additional exploration capability such as element identification and bio-chemical analysis. In the present report, principles, methods and specific research examples of CRM transmission radiography are summarized after which, principles, methods and perspective views of the future CAM radiography are described.

  11. Mesoscale Dynamical Regimes in the Midlatitudes

    NASA Astrophysics Data System (ADS)

    Craig, G. C.; Selz, T.

    2018-01-01

    The atmospheric mesoscales are characterized by a complex variety of meteorological phenomena that defy simple classification. Here a full space-time spectral analysis is carried out, based on a 7 day convection-permitting simulation of springtime midlatitude weather on a large domain. The kinetic energy is largest at synoptic scales, and on the mesoscale it is largely confined to an "advective band" where space and time scales are related by a constant of proportionality which corresponds to a velocity scale of about 10 m s-1. Computing the relative magnitude of different terms in the governing equations allows the identification of five dynamical regimes. These are tentatively identified as quasi-geostrophic flow, propagating gravity waves, stationary gravity waves related to orography, acoustic modes, and a weak temperature gradient regime, where vertical motions are forced by diabatic heating.

  12. ExprAlign - the identification of ESTs in non-model species by alignment of cDNA microarray expression profiles

    PubMed Central

    2009-01-01

    Background Sequence identification of ESTs from non-model species offers distinct challenges particularly when these species have duplicated genomes and when they are phylogenetically distant from sequenced model organisms. For the common carp, an environmental model of aquacultural interest, large numbers of ESTs remained unidentified using BLAST sequence alignment. We have used the expression profiles from large-scale microarray experiments to suggest gene identities. Results Expression profiles from ~700 cDNA microarrays describing responses of 7 major tissues to multiple environmental stressors were used to define a co-expression landscape. This was based on the Pearsons correlation coefficient relating each gene with all other genes, from which a network description provided clusters of highly correlated genes as 'mountains'. We show that these contain genes with known identities and genes with unknown identities, and that the correlation constitutes evidence of identity in the latter. This procedure has suggested identities to 522 of 2701 unknown carp ESTs sequences. We also discriminate several common carp genes and gene isoforms that were not discriminated by BLAST sequence alignment alone. Precision in identification was substantially improved by use of data from multiple tissues and treatments. Conclusion The detailed analysis of co-expression landscapes is a sensitive technique for suggesting an identity for the large number of BLAST unidentified cDNAs generated in EST projects. It is capable of detecting even subtle changes in expression profiles, and thereby of distinguishing genes with a common BLAST identity into different identities. It benefits from the use of multiple treatments or contrasts, and from the large-scale microarray data. PMID:19939286

  13. The OncoPPi Portal: an integrative resource to explore and prioritize protein-protein interactions for cancer target discovery. | Office of Cancer Genomics

    Cancer.gov

    Motivation: As cancer genomics initiatives move toward comprehensive identification of genetic alterations in cancer, attention is now turning to understanding how interactions among these genes lead to the acquisition of tumor hallmarks. Emerging pharmacological and clinical data suggest a highly promising role of cancer-specific protein-protein interactions (PPIs) as druggable cancer targets. However, large-scale experimental identification of cancer-related PPIs remains challenging, and currently available resources to explore oncogenic PPI networks are limited.

  14. AutoCNet: A Python library for sparse multi-image correspondence identification for planetary data

    NASA Astrophysics Data System (ADS)

    Laura, Jason; Rodriguez, Kelvin; Paquette, Adam C.; Dunn, Evin

    2018-01-01

    In this work we describe the AutoCNet library, written in Python, to support the application of computer vision techniques for n-image correspondence identification in remotely sensed planetary images and subsequent bundle adjustment. The library is designed to support exploratory data analysis, algorithm and processing pipeline development, and application at scale in High Performance Computing (HPC) environments for processing large data sets and generating foundational data products. We also present a brief case study illustrating high level usage for the Apollo 15 Metric camera.

  15. Structure identification methods for atomistic simulations of crystalline materials

    DOE PAGES

    Stukowski, Alexander

    2012-05-28

    Here, we discuss existing and new computational analysis techniques to classify local atomic arrangements in large-scale atomistic computer simulations of crystalline solids. This article includes a performance comparison of typical analysis algorithms such as common neighbor analysis (CNA), centrosymmetry analysis, bond angle analysis, bond order analysis and Voronoi analysis. In addition we propose a simple extension to the CNA method that makes it suitable for multi-phase systems. Finally, we introduce a new structure identification algorithm, the neighbor distance analysis, which is designed to identify atomic structure units in grain boundaries.

  16. bigSCale: an analytical framework for big-scale single-cell data.

    PubMed

    Iacono, Giovanni; Mereu, Elisabetta; Guillaumet-Adkins, Amy; Corominas, Roser; Cuscó, Ivon; Rodríguez-Esteban, Gustavo; Gut, Marta; Pérez-Jurado, Luis Alberto; Gut, Ivo; Heyn, Holger

    2018-06-01

    Single-cell RNA sequencing (scRNA-seq) has significantly deepened our insights into complex tissues, with the latest techniques capable of processing tens of thousands of cells simultaneously. Analyzing increasing numbers of cells, however, generates extremely large data sets, extending processing time and challenging computing resources. Current scRNA-seq analysis tools are not designed to interrogate large data sets and often lack sensitivity to identify marker genes. With bigSCale, we provide a scalable analytical framework to analyze millions of cells, which addresses the challenges associated with large data sets. To handle the noise and sparsity of scRNA-seq data, bigSCale uses large sample sizes to estimate an accurate numerical model of noise. The framework further includes modules for differential expression analysis, cell clustering, and marker identification. A directed convolution strategy allows processing of extremely large data sets, while preserving transcript information from individual cells. We evaluated the performance of bigSCale using both a biological model of aberrant gene expression in patient-derived neuronal progenitor cells and simulated data sets, which underlines the speed and accuracy in differential expression analysis. To test its applicability for large data sets, we applied bigSCale to assess 1.3 million cells from the mouse developing forebrain. Its directed down-sampling strategy accumulates information from single cells into index cell transcriptomes, thereby defining cellular clusters with improved resolution. Accordingly, index cell clusters identified rare populations, such as reelin ( Reln )-positive Cajal-Retzius neurons, for which we report previously unrecognized heterogeneity associated with distinct differentiation stages, spatial organization, and cellular function. Together, bigSCale presents a solution to address future challenges of large single-cell data sets. © 2018 Iacono et al.; Published by Cold Spring Harbor Laboratory Press.

  17. Identification of unknown apple cultivars demonstrates the impact of local breeding program on cultivar diversity

    USDA-ARS?s Scientific Manuscript database

    Apple trees, either abandoned or cared for, are common on the North American landscape. These trees can live for decades, and therefore represent a record of large- and small-scale agricultural practices through time. Here, we assessed the genetic diversity and identity of 330 unknown apple trees in...

  18. Identification of Preschool Children with Emotional Problems.

    ERIC Educational Resources Information Center

    Stern, Carolyn; And Others

    A large-scale study was designed to assess the extent of emotional disturbance among Head Start children and to provide a consistent basis for selection if therapeutic intervention were indicated. The study's aim was to avoid the problem of shifting baselines by individual teachers for determining the degree to which their children were departing…

  19. Design for a Study of American Youth.

    ERIC Educational Resources Information Center

    Flanagan, John C.; And Others

    Project TALENT is a large-scale, long-range educational research effort aimed at developing methods for the identification, development, and utilization of human talents, which has involved some 440,000 students in 1,353 public, private, and parochial secondary schools in all parts of the country. Data collected through teacher-administered tests,…

  20. Leveraging Technology to Improve Developmental Mathematics Course Completion: Evaluation of a Large-Scale Intervention

    ERIC Educational Resources Information Center

    Wladis, Claire; Offenholley, Kathleen; George, Michael

    2014-01-01

    This study hypothesizes that course passing rates in remedial mathematics classes can be improved through early identification of at-risk students using a department-wide midterm, followed by a mandated set of online intervention assignments incorporating immediate and elaborate feedback for all students identified as "at-risk" by their…

  1. Signs of the Times: Sources of Professional Challenge and Sustenance for Veteran African American Teachers

    ERIC Educational Resources Information Center

    Richards, Dean D., IV

    2017-01-01

    Recruitment and retention of high-quality educators remains problematic throughout our public school systems. This is particularly so for teachers of minority-identifications and in high-poverty, high-minority urban schools and districts. Recent research concerning teacher longevity has typically focused on large-scale investigations of factors of…

  2. Program Development: Identification and Formulation of Desirable Educational Goals.

    ERIC Educational Resources Information Center

    Goodlad, John I.

    In this speech, the author suggests that the success of public schools depends heavily on commitment to and large-scale agreement on educational goals. He examines the difficulty in creating rational programs to carry out specific behavioral goals and the more remote ends usually stated for educational systems. The author then discusses the…

  3. Identification of Student- and Teacher-Level Variables in Modeling Variation of Mathematics Achievement Data

    ERIC Educational Resources Information Center

    Tarr, James E.; Ross, Daniel J.; McNaught, Melissa D.; Chavez, Oscar; Grouws, Douglas A.; Reys, Robert E.; Sears, Ruthmae; Taylan, R. Didem

    2010-01-01

    The Comparing Options in Secondary Mathematics: Investigating Curriculum (COSMIC) project is a longitudinal study of student learning from two types of mathematics curricula: integrated and subject-specific. Previous large-scale research studies such as the National Assessment of Educational Progress (NAEP) indicate that numerous variables are…

  4. PepArML: A Meta-Search Peptide Identification Platform

    PubMed Central

    Edwards, Nathan J.

    2014-01-01

    The PepArML meta-search peptide identification platform provides a unified search interface to seven search engines; a robust cluster, grid, and cloud computing scheduler for large-scale searches; and an unsupervised, model-free, machine-learning-based result combiner, which selects the best peptide identification for each spectrum, estimates false-discovery rates, and outputs pepXML format identifications. The meta-search platform supports Mascot; Tandem with native, k-score, and s-score scoring; OMSSA; MyriMatch; and InsPecT with MS-GF spectral probability scores — reformatting spectral data and constructing search configurations for each search engine on the fly. The combiner selects the best peptide identification for each spectrum based on search engine results and features that model enzymatic digestion, retention time, precursor isotope clusters, mass accuracy, and proteotypic peptide properties, requiring no prior knowledge of feature utility or weighting. The PepArML meta-search peptide identification platform often identifies 2–3 times more spectra than individual search engines at 10% FDR. PMID:25663956

  5. Quality Assessments of Long-Term Quantitative Proteomic Analysis of Breast Cancer Xenograft Tissues

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhou, Jian-Ying; Chen, Lijun; Zhang, Bai

    The identification of protein biomarkers requires large-scale analysis of human specimens to achieve statistical significance. In this study, we evaluated the long-term reproducibility of an iTRAQ (isobaric tags for relative and absolute quantification) based quantitative proteomics strategy using one channel for universal normalization across all samples. A total of 307 liquid chromatography tandem mass spectrometric (LC-MS/MS) analyses were completed, generating 107 one-dimensional (1D) LC-MS/MS datasets and 8 offline two-dimensional (2D) LC-MS/MS datasets (25 fractions for each set) for human-in-mouse breast cancer xenograft tissues representative of basal and luminal subtypes. Such large-scale studies require the implementation of robust metrics to assessmore » the contributions of technical and biological variability in the qualitative and quantitative data. Accordingly, we developed a quantification confidence score based on the quality of each peptide-spectrum match (PSM) to remove quantification outliers from each analysis. After combining confidence score filtering and statistical analysis, reproducible protein identification and quantitative results were achieved from LC-MS/MS datasets collected over a 16 month period.« less

  6. Systems identification and the adaptive management of waterfowl in the United States

    USGS Publications Warehouse

    Williams, B.K.; Nichols, J.D.

    2001-01-01

    Waterfowl management in the United States is one of the more visible conservation success stories in the United States. It is authorized and supported by appropriate legislative authorities, based on large-scale monitoring programs, and widely accepted by the public. The process is one of only a limited number of large-scale examples of effective collaboration between research and management, integrating scientific information with management in a coherent framework for regulatory decision-making. However, harvest management continues to face some serious technical problems, many of which focus on sequential identification of the resource system in a context of optimal decision-making. The objective of this paper is to provide a theoretical foundation of adaptive harvest management, the approach currently in use in the United States for regulatory decision-making. We lay out the legal and institutional framework for adaptive harvest management and provide a formal description of regulatory decision-making in terms of adaptive optimization. We discuss some technical and institutional challenges in applying adaptive harvest management and focus specifically on methods of estimating resource states for linear resource systems.

  7. Multi-color electron microscopy by element-guided identification of cells, organelles and molecules.

    PubMed

    Scotuzzi, Marijke; Kuipers, Jeroen; Wensveen, Dasha I; de Boer, Pascal; Hagen, Kees C W; Hoogenboom, Jacob P; Giepmans, Ben N G

    2017-04-07

    Cellular complexity is unraveled at nanometer resolution using electron microscopy (EM), but interpretation of macromolecular functionality is hampered by the difficulty in interpreting grey-scale images and the unidentified molecular content. We perform large-scale EM on mammalian tissue complemented with energy-dispersive X-ray analysis (EDX) to allow EM-data analysis based on elemental composition. Endogenous elements, labels (gold and cadmium-based nanoparticles) as well as stains are analyzed at ultrastructural resolution. This provides a wide palette of colors to paint the traditional grey-scale EM images for composition-based interpretation. Our proof-of-principle application of EM-EDX reveals that endocrine and exocrine vesicles exist in single cells in Islets of Langerhans. This highlights how elemental mapping reveals unbiased biomedical relevant information. Broad application of EM-EDX will further allow experimental analysis on large-scale tissue using endogenous elements, multiple stains, and multiple markers and thus brings nanometer-scale 'color-EM' as a promising tool to unravel molecular (de)regulation in biomedicine.

  8. Multi-color electron microscopy by element-guided identification of cells, organelles and molecules

    PubMed Central

    Scotuzzi, Marijke; Kuipers, Jeroen; Wensveen, Dasha I.; de Boer, Pascal; Hagen, Kees (C.) W.; Hoogenboom, Jacob P.; Giepmans, Ben N. G.

    2017-01-01

    Cellular complexity is unraveled at nanometer resolution using electron microscopy (EM), but interpretation of macromolecular functionality is hampered by the difficulty in interpreting grey-scale images and the unidentified molecular content. We perform large-scale EM on mammalian tissue complemented with energy-dispersive X-ray analysis (EDX) to allow EM-data analysis based on elemental composition. Endogenous elements, labels (gold and cadmium-based nanoparticles) as well as stains are analyzed at ultrastructural resolution. This provides a wide palette of colors to paint the traditional grey-scale EM images for composition-based interpretation. Our proof-of-principle application of EM-EDX reveals that endocrine and exocrine vesicles exist in single cells in Islets of Langerhans. This highlights how elemental mapping reveals unbiased biomedical relevant information. Broad application of EM-EDX will further allow experimental analysis on large-scale tissue using endogenous elements, multiple stains, and multiple markers and thus brings nanometer-scale ‘color-EM’ as a promising tool to unravel molecular (de)regulation in biomedicine. PMID:28387351

  9. Finding a needle in the virus metagenome haystack--micro-metagenome analysis captures a snapshot of the diversity of a bacteriophage armoire.

    PubMed

    Ray, Jessica; Dondrup, Michael; Modha, Sejal; Steen, Ida Helene; Sandaa, Ruth-Anne; Clokie, Martha

    2012-01-01

    Viruses are ubiquitous in the oceans and critical components of marine microbial communities, regulating nutrient transfer to higher trophic levels or to the dissolved organic pool through lysis of host cells. Hydrothermal vent systems are oases of biological activity in the deep oceans, for which knowledge of biodiversity and its impact on global ocean biogeochemical cycling is still in its infancy. In order to gain biological insight into viral communities present in hydrothermal vent systems, we developed a method based on deep-sequencing of pulsed field gel electrophoretic bands representing key viral fractions present in seawater within and surrounding a hydrothermal plume derived from Loki's Castle vent field at the Arctic Mid-Ocean Ridge. The reduction in virus community complexity afforded by this novel approach enabled the near-complete reconstruction of a lambda-like phage genome from the virus fraction of the plume. Phylogenetic examination of distinct gene regions in this lambdoid phage genome unveiled diversity at loci encoding superinfection exclusion- and integrase-like proteins. This suggests the importance of fine-tuning lyosgenic conversion as a viral survival strategy, and provides insights into the nature of host-virus and virus-virus interactions, within hydrothermal plumes. By reducing the complexity of the viral community through targeted sequencing of prominent dsDNA viral fractions, this method has selectively mimicked virus dominance approaching that hitherto achieved only through culturing, thus enabling bioinformatic analysis to locate a lambdoid viral "needle" within the greater viral community "haystack". Such targeted analyses have great potential for accelerating the extraction of biological knowledge from diverse and poorly understood environmental viral communities.

  10. Monstrous Ice Cloud System in Titan's Present South Polar Stratosphere

    NASA Astrophysics Data System (ADS)

    Anderson, Carrie; Samuelson, Robert; McLain, Jason; Achterberg, Richard; Flasar, F. Michael; Milam, Stefanie

    2015-11-01

    During southern autumn when sunlight was still available, Cassini's Imaging Science Subsystem discovered a cloud around 300 km near Titan's south pole (West, R. A. et al., AAS/DPS Abstracts, 45, #305.03, 2013); the cloud was later determined by Cassini's Visible and InfraRed Mapping Spectrometer to contain HCN ice (de Kok et al., Nature, 514, pp 65-67, 2014). This cloud has proven to be only the tip of an extensive ice cloud system contained in Titan's south polar stratosphere, as seen through the night-vision goggles of Cassini's Composite InfraRed Spectrometer (CIRS). As the sun sets and the gloom of southern winter approaches, evidence is beginning to accumulate from CIRS far-IR spectra that a massive system of nitrile ice clouds is developing in Titan's south polar stratosphere. Even during the depths of northern winter, nothing like the strength of this southern system was evident in corresponding north polar regions.From the long slant paths that are available from limb-viewing CIRS far-IR spectra, we have the first definitive detection of the ν6 band of cyanoacetylene (HC3N) ice in Titan’s south polar stratosphere. In addition, we also see a strong blend of nitrile ice lattice vibration features around 160 cm-1. From these data we are able to derive ice abundances. The most prominent (and still chemically unidentified) ice emission feature, the Haystack, (at 220 cm-1) is also observed. We establish the vertical distributions of the ice cloud systems associated with both the 160 cm-1 feature and the Haystack. The ultimate aim is to refine the physical and possibly the chemical relationships between the two. Transmittance thin film spectra of nitrile ice mixtures obtained in our Spectroscopy for Planetary ICes Environments (SPICE) laboratory are used to support these analyses.

  11. An Investigation of the Large Scale Evolution and Topology of Coronal Mass Ejections in the Solar Wind

    NASA Technical Reports Server (NTRS)

    Riley, Peter

    2000-01-01

    This investigation is concerned with the large-scale evolution and topology of coronal mass ejections (CMEs) in the solar wind. During this reporting period we have focused on several aspects of CME properties, their identification and their evolution in the solar wind. The work included both analysis of Ulysses and ACE observations as well as fluid and magnetohydrodynamic simulations. In addition, we analyzed a series of "density holes" observed in the solar wind, that bear many similarities with CMEs. Finally, this work was communicated to the scientific community at three meetings and has led to three scientific papers that are in various stages of review.

  12. Efficient estimation and large-scale evaluation of lateral chromatic aberration for digital image forensics

    NASA Astrophysics Data System (ADS)

    Gloe, Thomas; Borowka, Karsten; Winkler, Antje

    2010-01-01

    The analysis of lateral chromatic aberration forms another ingredient for a well equipped toolbox of an image forensic investigator. Previous work proposed its application to forgery detection1 and image source identification.2 This paper takes a closer look on the current state-of-the-art method to analyse lateral chromatic aberration and presents a new approach to estimate lateral chromatic aberration in a runtime-efficient way. Employing a set of 11 different camera models including 43 devices, the characteristic of lateral chromatic aberration is investigated in a large-scale. The reported results point to general difficulties that have to be considered in real world investigations.

  13. Near-optimal quantum circuit for Grover's unstructured search using a transverse field

    NASA Astrophysics Data System (ADS)

    Jiang, Zhang; Rieffel, Eleanor G.; Wang, Zhihui

    2017-06-01

    Inspired by a class of algorithms proposed by Farhi et al. (arXiv:1411.4028), namely, the quantum approximate optimization algorithm (QAOA), we present a circuit-based quantum algorithm to search for a needle in a haystack, obtaining the same quadratic speedup achieved by Grover's original algorithm. In our algorithm, the problem Hamiltonian (oracle) and a transverse field are applied alternately to the system in a periodic manner. We introduce a technique, based on spin-coherent states, to analyze the composite unitary in a single period. This composite unitary drives a closed transition between two states that have high degrees of overlap with the initial state and the target state, respectively. The transition rate in our algorithm is of order Θ (1 /√{N }) , and the overlaps are of order Θ (1 ) , yielding a nearly optimal query complexity of T ≃√{N }(π /2 √{2 }) . Our algorithm is a QAOA circuit that demonstrates a quantum advantage with a large number of iterations that is not derived from Trotterization of an adiabatic quantum optimization (AQO) algorithm. It also suggests that the analysis required to understand QAOA circuits involves a very different process from estimating the energy gap of a Hamiltonian in AQO.

  14. The Characteristics and Consequences of the Break-up of the Fengyun-1C Spacecraft

    NASA Technical Reports Server (NTRS)

    Johnson, Nicholas L.; Stansbery, Eugene; Liou, Jer-chyi; Horstman, Matt; Stokeley, Christopher; Whitlock, David

    2007-01-01

    The intentional break-up of the Fengyun-1C spacecraft on 11 January 2007 via hypervelocity collision with a ballistic object created the most severe artificial debris cloud in Earth orbit since the beginning of space exploration. More than 900 debris on the order of 10 cm or greater in size have been identified by the U.S. Space Surveillance Network (SSN). The majority of these debris reside in long-lived orbits. The NASA Orbital Debris Program Office has conducted a thorough examination of the nature of the Fengyun-1C debris cloud, using SSN data for larger debris and special Haystack radar observations for smaller debris. These data have been compared with the NASA standard satellite break-up model for collisions, and the results are presented in this paper. The orbital longevity of the debris have also been evaluated for both small and large debris. The consequent long-term spatial density effects on the low Earth orbit (LEO) regime are then described. Finally, collision probabilities between the Fengyun-1C debris cloud and the resident space object population of 1 January 2007 have been calculated. The potential effect on the growth of the near-Earth satellite population is presented.

  15. Identification of Stevioside Using Tissue Culture-Derived Stevia (Stevia rebaudiana) Leaves

    PubMed Central

    Karim, Md. Ziaul; Uesugi, Daisuke; Nakayama, Noriyuki; Hossain, M. Monzur; Ishihara, Kohji; Hamada, Hiroki

    2015-01-01

    Stevioside is a natural sweetener from Stevia leaf, which is 300 times sweeter than sugar. It helps to reduce blood sugar levels dramatically and thus can be of benefit to diabetic people. Tissue culture is a very potential modern technology that can be used in large-scale disease-free stevia production throughout the year. We successfully produced stevia plant through in vitro culture for identification of stevioside in this experiment. The present study describes a potential method for identification of stevioside from tissue culture-derived stevia leaf. Stevioside in the sample was identified using HPLC by measuring the retention time. The percentage of stevioside content in the leaf samples was found to be 9.6%. This identification method can be used for commercial production and industrialization of stevia through in vitro culture across the world. PMID:28008268

  16. Mapping ecosystem services for land use planning, the case of Central Kalimantan.

    PubMed

    Sumarga, Elham; Hein, Lars

    2014-07-01

    Indonesia is subject to rapid land use change. One of the main causes for the conversion of land is the rapid expansion of the oil palm sector. Land use change involves a progressive loss of forest cover, with major impacts on biodiversity and global CO2 emissions. Ecosystem services have been proposed as a concept that would facilitate the identification of sustainable land management options, however, the scale of land conversion and its spatial diversity pose particular challenges in Indonesia. The objective of this paper is to analyze how ecosystem services can be mapped at the provincial scale, focusing on Central Kalimantan, and to examine how ecosystem services maps can be used for a land use planning. Central Kalimantan is subject to rapid deforestation including the loss of peatland forests and the provincial still lacks a comprehensive land use plan. We examine how seven key ecosystem services can be mapped and modeled at the provincial scale, using a variety of models, and how large scale ecosystem services maps can support the identification of options for sustainable expansion of palm oil production.

  17. Isolation and characterizations of oxalate-binding proteins in the kidney

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Roop-ngam, Piyachat; Chaiyarit, Sakdithep; Pongsakul, Nutkridta

    Highlights: Black-Right-Pointing-Pointer The first large-scale characterizations of oxalate-binding kidney proteins. Black-Right-Pointing-Pointer The recently developed oxalate-conjugated EAH Sepharose 4B beads were applied. Black-Right-Pointing-Pointer 38 forms of 26 unique oxalate-binding kidney proteins were identified. Black-Right-Pointing-Pointer 25/26 (96%) of identified proteins had 'L-x(3,5)-R-x(2)-[AGILPV]' domain. -- Abstract: Oxalate-binding proteins are thought to serve as potential modulators of kidney stone formation. However, only few oxalate-binding proteins have been identified from previous studies. Our present study, therefore, aimed for large-scale identification of oxalate-binding proteins in porcine kidney using an oxalate-affinity column containing oxalate-conjugated EAH Sepharose 4B beads for purification followed by two-dimensional gel electrophoresis (2-DE) tomore » resolve the recovered proteins. Comparing with those obtained from the controlled column containing uncoupled EAH-Sepharose 4B (to subtract the background of non-specific bindings), a total of 38 protein spots were defined as oxalate-binding proteins. These protein spots were successfully identified by quadrupole time-of-flight mass spectrometry (MS) and/or tandem MS (MS/MS) as 26 unique proteins, including several nuclear proteins, mitochondrial proteins, oxidative stress regulatory proteins, metabolic enzymes and others. Identification of oxalate-binding domain using the PRATT tool revealed 'L-x(3,5)-R-x(2)-[AGILPV]' as a functional domain responsible for oxalate-binding in 25 of 26 (96%) unique identified proteins. We report herein, for the first time, large-scale identification and characterizations of oxalate-binding proteins in the kidney. The presence of positively charged arginine residue in the middle of this functional domain suggested its significance for binding to the negatively charged oxalate. These data will enhance future stone research, particularly on stone modulators.« less

  18. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Larkoski, Andrew J.; Maltoni, Fabio; Selvaggi, Michele

    The identification of hadronically decaying heavy states, such as vector bosons, the Higgs, or the top quark, produced with large transverse boosts has been and will continue to be a central focus of the jet physics program at the Large Hadron Collider (LHC). At a future hadron collider working at an order-of-magnitude larger energy than the LHC, these heavy states would be easily produced with transverse boosts of several TeV. At these energies, their decay products will be separated by angular scales comparable to individual calorimeter cells, making the current jet substructure identification techniques for hadronic decay modes not directlymore » employable. In addition, at the high energy and luminosity projected at a future hadron collider, there will be numerous sources for contamination including initial- and final-state radiation, underlying event, or pile-up which must be mitigated. We propose a simple strategy to tag such "hyper-boosted" objects that defines jets with radii that scale inversely proportional to their transverse boost and combines the standard calorimetric information with charged track-based observables. By means of a fast detector simulation, we apply it to top quark identification and demonstrate that our method efficiently discriminates hadronically decaying top quarks from light QCD jets up to transverse boosts of 20 TeV. Lastly, our results open the way to tagging heavy objects with energies in the multi-TeV range at present and future hadron colliders.« less

  19. A Novel Proteomics Approach to Identify SUMOylated Proteins and Their Modification Sites in Human Cells*

    PubMed Central

    Galisson, Frederic; Mahrouche, Louiza; Courcelles, Mathieu; Bonneil, Eric; Meloche, Sylvain; Chelbi-Alix, Mounira K.; Thibault, Pierre

    2011-01-01

    The small ubiquitin-related modifier (SUMO) is a small group of proteins that are reversibly attached to protein substrates to modify their functions. The large scale identification of protein SUMOylation and their modification sites in mammalian cells represents a significant challenge because of the relatively small number of in vivo substrates and the dynamic nature of this modification. We report here a novel proteomics approach to selectively enrich and identify SUMO conjugates from human cells. We stably expressed different SUMO paralogs in HEK293 cells, each containing a His6 tag and a strategically located tryptic cleavage site at the C terminus to facilitate the recovery and identification of SUMOylated peptides by affinity enrichment and mass spectrometry. Tryptic peptides with short SUMO remnants offer significant advantages in large scale SUMOylome experiments including the generation of paralog-specific fragment ions following CID and ETD activation, and the identification of modified peptides using conventional database search engines such as Mascot. We identified 205 unique protein substrates together with 17 precise SUMOylation sites present in 12 SUMO protein conjugates including three new sites (Lys-380, Lys-400, and Lys-497) on the protein promyelocytic leukemia. Label-free quantitative proteomics analyses on purified nuclear extracts from untreated and arsenic trioxide-treated cells revealed that all identified SUMOylated sites of promyelocytic leukemia were differentially SUMOylated upon stimulation. PMID:21098080

  20. Tracking down hyper-boosted top quarks

    DOE PAGES

    Larkoski, Andrew J.; Maltoni, Fabio; Selvaggi, Michele

    2015-06-05

    The identification of hadronically decaying heavy states, such as vector bosons, the Higgs, or the top quark, produced with large transverse boosts has been and will continue to be a central focus of the jet physics program at the Large Hadron Collider (LHC). At a future hadron collider working at an order-of-magnitude larger energy than the LHC, these heavy states would be easily produced with transverse boosts of several TeV. At these energies, their decay products will be separated by angular scales comparable to individual calorimeter cells, making the current jet substructure identification techniques for hadronic decay modes not directlymore » employable. In addition, at the high energy and luminosity projected at a future hadron collider, there will be numerous sources for contamination including initial- and final-state radiation, underlying event, or pile-up which must be mitigated. We propose a simple strategy to tag such "hyper-boosted" objects that defines jets with radii that scale inversely proportional to their transverse boost and combines the standard calorimetric information with charged track-based observables. By means of a fast detector simulation, we apply it to top quark identification and demonstrate that our method efficiently discriminates hadronically decaying top quarks from light QCD jets up to transverse boosts of 20 TeV. Lastly, our results open the way to tagging heavy objects with energies in the multi-TeV range at present and future hadron colliders.« less

  1. Application of stochastic models in identification and apportionment of heavy metal pollution sources in the surface soils of a large-scale region.

    PubMed

    Hu, Yuanan; Cheng, Hefa

    2013-04-16

    As heavy metals occur naturally in soils at measurable concentrations and their natural background contents have significant spatial variations, identification and apportionment of heavy metal pollution sources across large-scale regions is a challenging task. Stochastic models, including the recently developed conditional inference tree (CIT) and the finite mixture distribution model (FMDM), were applied to identify the sources of heavy metals found in the surface soils of the Pearl River Delta, China, and to apportion the contributions from natural background and human activities. Regression trees were successfully developed for the concentrations of Cd, Cu, Zn, Pb, Cr, Ni, As, and Hg in 227 soil samples from a region of over 7.2 × 10(4) km(2) based on seven specific predictors relevant to the source and behavior of heavy metals: land use, soil type, soil organic carbon content, population density, gross domestic product per capita, and the lengths and classes of the roads surrounding the sampling sites. The CIT and FMDM results consistently indicate that Cd, Zn, Cu, Pb, and Cr in the surface soils of the PRD were contributed largely by anthropogenic sources, whereas As, Ni, and Hg in the surface soils mostly originated from the soil parent materials.

  2. Single-shot stand-off chemical identification of powders using random Raman lasing

    PubMed Central

    Hokr, Brett H.; Bixler, Joel N.; Noojin, Gary D.; Thomas, Robert J.; Rockwell, Benjamin A.; Yakovlev, Vladislav V.; Scully, Marlan O.

    2014-01-01

    The task of identifying explosives, hazardous chemicals, and biological materials from a safe distance is the subject we consider. Much of the prior work on stand-off spectroscopy using light has been devoted to generating a backward-propagating beam of light that can be used drive further spectroscopic processes. The discovery of random lasing and, more recently, random Raman lasing provide a mechanism for remotely generating copious amounts of chemically specific Raman scattered light. The bright nature of random Raman lasing renders directionality unnecessary, allowing for the detection and identification of chemicals from large distances in real time. In this article, the single-shot remote identification of chemicals at kilometer-scale distances is experimentally demonstrated using random Raman lasing. PMID:25114231

  3. Materials identification using a small-scale pixellated x-ray diffraction system

    NASA Astrophysics Data System (ADS)

    O'Flynn, D.; Crews, C.; Drakos, I.; Christodoulou, C.; Wilson, M. D.; Veale, M. C.; Seller, P.; Speller, R. D.

    2016-05-01

    A transmission x-ray diffraction system has been developed using a pixellated, energy-resolving detector (HEXITEC) and a small-scale, mains operated x-ray source (Amptek Mini-X). HEXITEC enables diffraction to be measured without the requirement of incident spectrum filtration, or collimation of the scatter from the sample, preserving a large proportion of the useful signal compared with other diffraction techniques. Due to this efficiency, sufficient molecular information for material identification can be obtained within 5 s despite the relatively low x-ray source power. Diffraction data are presented from caffeine, hexamine, paracetamol, plastic explosives and narcotics. The capability to determine molecular information from aspirin tablets inside their packaging is demonstrated. Material selectivity and the potential for a sample classification model is shown with principal component analysis, through which each different material can be clearly resolved.

  4. The benefits of using remotely sensed soil moisture in parameter identification of large-scale hydrological models

    NASA Astrophysics Data System (ADS)

    Wanders, N.; Bierkens, M. F. P.; de Jong, S. M.; de Roo, A.; Karssenberg, D.

    2014-08-01

    Large-scale hydrological models are nowadays mostly calibrated using observed discharge. As a result, a large part of the hydrological system, in particular the unsaturated zone, remains uncalibrated. Soil moisture observations from satellites have the potential to fill this gap. Here we evaluate the added value of remotely sensed soil moisture in calibration of large-scale hydrological models by addressing two research questions: (1) Which parameters of hydrological models can be identified by calibration with remotely sensed soil moisture? (2) Does calibration with remotely sensed soil moisture lead to an improved calibration of hydrological models compared to calibration based only on discharge observations, such that this leads to improved simulations of soil moisture content and discharge? A dual state and parameter Ensemble Kalman Filter is used to calibrate the hydrological model LISFLOOD for the Upper Danube. Calibration is done using discharge and remotely sensed soil moisture acquired by AMSR-E, SMOS, and ASCAT. Calibration with discharge data improves the estimation of groundwater and routing parameters. Calibration with only remotely sensed soil moisture results in an accurate identification of parameters related to land-surface processes. For the Upper Danube upstream area up to 40,000 km2, calibration on both discharge and soil moisture results in a reduction by 10-30% in the RMSE for discharge simulations, compared to calibration on discharge alone. The conclusion is that remotely sensed soil moisture holds potential for calibration of hydrological models, leading to a better simulation of soil moisture content throughout the catchment and a better simulation of discharge in upstream areas. This article was corrected on 15 SEP 2014. See the end of the full text for details.

  5. Attaining insight into interactions between hydrologic model parameters and geophysical attributes for national-scale model parameter estimation

    NASA Astrophysics Data System (ADS)

    Mizukami, N.; Clark, M. P.; Newman, A. J.; Wood, A.; Gutmann, E. D.

    2017-12-01

    Estimating spatially distributed model parameters is a grand challenge for large domain hydrologic modeling, especially in the context of hydrologic model applications such as streamflow forecasting. Multi-scale Parameter Regionalization (MPR) is a promising technique that accounts for the effects of fine-scale geophysical attributes (e.g., soil texture, land cover, topography, climate) on model parameters and nonlinear scaling effects on model parameters. MPR computes model parameters with transfer functions (TFs) that relate geophysical attributes to model parameters at the native input data resolution and then scales them using scaling functions to the spatial resolution of the model implementation. One of the biggest challenges in the use of MPR is identification of TFs for each model parameter: both functional forms and geophysical predictors. TFs used to estimate the parameters of hydrologic models typically rely on previous studies or were derived in an ad-hoc, heuristic manner, potentially not utilizing maximum information content contained in the geophysical attributes for optimal parameter identification. Thus, it is necessary to first uncover relationships among geophysical attributes, model parameters, and hydrologic processes (i.e., hydrologic signatures) to obtain insight into which and to what extent geophysical attributes are related to model parameters. We perform multivariate statistical analysis on a large-sample catchment data set including various geophysical attributes as well as constrained VIC model parameters at 671 unimpaired basins over the CONUS. We first calibrate VIC model at each catchment to obtain constrained parameter sets. Additionally, parameter sets sampled during the calibration process are used for sensitivity analysis using various hydrologic signatures as objectives to understand the relationships among geophysical attributes, parameters, and hydrologic processes.

  6. A novel strategy for global mapping of O-GlcNAc proteins and peptides using selective enzymatic deglycosylation, HILIC enrichment and mass spectrometry identification.

    PubMed

    Shen, Bingquan; Zhang, Wanjun; Shi, Zhaomei; Tian, Fang; Deng, Yulin; Sun, Changqing; Wang, Guangshun; Qin, Weijie; Qian, Xiaohong

    2017-07-01

    O-GlcNAcylation is a kind of dynamic O-linked glycosylation of nucleocytoplasmic and mitochondrial proteins. It serves as a major nutrient sensor to regulate numerous biological processes including transcriptional regulation, cell metabolism, cellular signaling, and protein degradation. Dysregulation of cellular O-GlcNAcylated levels contributes to the etiologies of many diseases such as diabetes, neurodegenerative disease and cancer. However, deeper insight into the biological mechanism of O-GlcNAcylation is hampered by its extremely low stoichiometry and the lack of efficient enrichment approaches for large-scale identification by mass spectrometry. Herein, we developed a novel strategy for the global identification of O-GlcNAc proteins and peptides using selective enzymatic deglycosylation, HILIC enrichment and mass spectrometry analysis. Standard O-GlcNAc peptides can be efficiently enriched even in the presence of 500-fold more abundant non-O-GlcNAc peptides and identified by mass spectrometry with a low nanogram detection sensitivity. This strategy successfully achieved the first large-scale enrichment and characterization of O-GlcNAc proteins and peptides in human urine. A total of 474 O-GlcNAc peptides corresponding to 457 O-GlcNAc proteins were identified by mass spectrometry analysis, which is at least three times more than that obtained by commonly used enrichment methods. A large number of unreported O-GlcNAc proteins related to cell cycle, biological regulation, metabolic and developmental process were found in our data. The above results demonstrated that this novel strategy is highly efficient in the global enrichment and identification of O-GlcNAc peptides. These data provide new insights into the biological function of O-GlcNAcylation in human urine, which is correlated with the physiological states and pathological changes of human body and therefore indicate the potential of this strategy for biomarker discovery from human urine. Copyright © 2017. Published by Elsevier B.V.

  7. Demystifying Scientific Data ­ Using Earth Science to Teach the Scientific Method

    NASA Astrophysics Data System (ADS)

    Nassiff, P. J.; Santos, E. A.; Erickson, P. J.; Niell, A. E.

    2006-12-01

    The collection of large quantities of data and their subsequent analyses are important components of any scientific process, particularly at research institutes such as MIT's Haystack Observatory, where the collection and analyses of data is crucial to research efforts. Likewise, a recent study on science education concluded that students should be introduced to analyzing evidence and hypotheses, to critical thinking - including appropriate skepticism, to quantitative reasoning and the ability to make reasonable estimates, and to the role of uncertainty and error in science. In order to achieve this goal with grades 9-12 students and their instructors, we developed lesson plans and activities based on atmospheric science and geodetic research at Haystack Observatory. From the complex steps of experimental design, measurement, and data analysis, students and teachers will gain insight into the scientific research processes as they exist today. The use of these space weather and geodesy activities in classrooms will be discussed. Space Weather: After decades of data collection with multiple variables, space weather is about as complex an area of investigation as possible. Far from the passive relationship between the Sun and Earth often taught in the early grades, or the beautiful auroras discussed in high school, there are complex and powerful interactions between the Sun and Earth. In spite of these complexities, high school students can learn about space weather and the repercussions on our communication and power technologies. Starting from lessons on the basic method of observing space weather with incoherent scatter radar, and progressing to the use of simplified data sets, students will discover how space weather affects Earth over solar cycles and how severe solar activity is measured and affects the Earth over shorter time spans. They will see that even from complex, seemingly ambiguous data with many variables and unknowns, scientists can gain valuable insights into complicated processes. Geodesy: Students learn about tectonic plate theory in middle school to explain continental drift, but have no idea about how it is determined. By learning about the process, students become more familiar with measurement, uncertainty, and error. Students who analyze continental drift using observations from very long baseline interferometry (VLBI) will discover the current limits of scientific measurement (approximately one part in a billion) and see how even these data may contain unmodeled effects. In both projects the process of understanding data will give the students a better picture of how science works. These lessons and activities were created under the Research Experiences for Teachers program of the National Science Foundation.

  8. Environmental Sensitivity in Children: Development of the Highly Sensitive Child Scale and Identification of Sensitivity Groups

    ERIC Educational Resources Information Center

    Pluess, Michael; Assary, Elham; Lionetti, Francesca; Lester, Kathryn J.; Krapohl, Eva; Aron, Elaine N.; Aron, Arthur

    2018-01-01

    A large number of studies document that children differ in the degree they are shaped by their developmental context with some being more sensitive to environmental influences than others. Multiple theories suggest that "Environmental Sensitivity" is a common trait predicting the response to negative as well as positive exposures.…

  9. Fusion of Remote Sensing and Non-Authoritative Data for Flood Disaster and Transportation Infrastructure Assessment

    ERIC Educational Resources Information Center

    Schnebele, Emily K.

    2013-01-01

    Flooding is the most frequently occurring natural hazard on Earth; with catastrophic, large scale floods causing immense damage to people, property, and the environment. Over the past 20 years, remote sensing has become the standard technique for flood identification because of its ability to offer synoptic coverage. Unfortunately, remote sensing…

  10. The UAB Informatics Institute and 2016 CEGS N-GRID de-identification shared task challenge.

    PubMed

    Bui, Duy Duc An; Wyatt, Mathew; Cimino, James J

    2017-11-01

    Clinical narratives (the text notes found in patients' medical records) are important information sources for secondary use in research. However, in order to protect patient privacy, they must be de-identified prior to use. Manual de-identification is considered to be the gold standard approach but is tedious, expensive, slow, and impractical for use with large-scale clinical data. Automated or semi-automated de-identification using computer algorithms is a potentially promising alternative. The Informatics Institute of the University of Alabama at Birmingham is applying de-identification to clinical data drawn from the UAB hospital's electronic medical records system before releasing them for research. We participated in a shared task challenge by the Centers of Excellence in Genomic Science (CEGS) Neuropsychiatric Genome-Scale and RDoC Individualized Domains (N-GRID) at the de-identification regular track to gain experience developing our own automatic de-identification tool. We focused on the popular and successful methods from previous challenges: rule-based, dictionary-matching, and machine-learning approaches. We also explored new techniques such as disambiguation rules, term ambiguity measurement, and used multi-pass sieve framework at a micro level. For the challenge's primary measure (strict entity), our submissions achieved competitive results (f-measures: 87.3%, 87.1%, and 86.7%). For our preferred measure (binary token HIPAA), our submissions achieved superior results (f-measures: 93.7%, 93.6%, and 93%). With those encouraging results, we gain the confidence to improve and use the tool for the real de-identification task at the UAB Informatics Institute. Copyright © 2017 Elsevier Inc. All rights reserved.

  11. Search for Genetic Modifiers of PSC: Time to Increase the Number of Needles in the Haystack.

    PubMed

    Krawczyk, Marcin; Lammert, Frank

    Primary sclerosing cholangitis (PSC) belongs to the most obscure liver diseases. Patients with progressive PSC require liver transplantation as only therapeutic option. Previously several HLA- and non-HLA-associated PSC risk variants have been discovered, however their involvement in the development of PSC seems to be minor in comparison to environmental determinants. Lately, variant rs853974 at the RSPO3 gene locus has been shown to modulate the course of PSC. Here we briefly discuss the phenotypes related to this polymorphism and propose alternative directions of research that might help to identify new genetic modifiers of PSC progression.

  12. RDBE Development and Progress

    NASA Astrophysics Data System (ADS)

    Neill, A.; Bark, M.; Beaudoin, C.; Brisken, W.; Ben Frej, H.; Doeleman, S.; Durand, S.; Guerra, Ml; Hinton, A.; Luce, M.; McWhirter, R.; Morris, K.; Peck, G.; Revnell, M.; Rogers, A.; Romney, J.; Ruszczyk, C; Taveniku, M.; Walker, R.; Whitney, A.

    2010-12-01

    A digital backend based on the ROACH board has been developed jointly by the National Radio Astronomy Observatory and MIT Haystack Observatory. The RDBE will have both Polyphase Filterbank and Digital Downconverter personalities. The initial configuration outputs sixteen 32-MHz channels, comprised of half the channels from the PFB processing of the two IF inputs, for use in the VLBI2010 geodetic system and in the VLBA sensitivity upgrade project. The output rate is 2x10^9 bits/second (1x10^9 bits/sec = 1 Gbps) over a 10 GigE connection to the Mark 5C with the data written in Mark 5B format on disk.

  13. RDBE Development and Progress

    NASA Technical Reports Server (NTRS)

    Niell, A.; Bark, M.; Beaudoin, C.; Brisken, W.; Frej, H. Ben; Doeleman, S.; Durand, S.; Guerra, M.; Hinton, A.; Luce, M.; hide

    2010-01-01

    A digital backend based on the ROACH board has been developed jointly by the National Radio Astronomy Observatory and MIT Haystack Observatory. The RDBE will have both Polyphase Filterbank and Digital Downconverter personalities. The initial configuration outputs sixteen 32-MHz channels, comprised of half the channels from the PFB processing of the two IF inputs, for use in the VLBI2010 geodetic system and in the VLBA sensitivity upgrade project. The output rate is 2x109 bits/second (1x10(exp 9) bits/sec = 1 Gbps) over a 10 GigE connection to the Mark 5C with the data written in Mark 5B format on disk.

  14. Body identification, biometrics and medicine: ethical and social considerations.

    PubMed

    Mordini, Emilio; Ottolini, Corinna

    2007-01-01

    Identity is important when it is weak. This apparent paradox is the core of the current debate on identity. Traditionally, verification of identity has been based upon authentication of attributed and biographical characteristics. After small scale societies and large scale, industrial societies, globalization represents the third period of personal identification. The human body lies at the heart of all strategies for identity management. The tension between human body and personal identity is critical in the health care sector. The health care sector is second only to the financial sector in term of the number of biometric users. Many hospitals and healthcare organizations are in progress to deploy biometric security architecture. Secure identification is critical in the health care system, both to control logic access to centralized archives of digitized patients' data, and to limit physical access to buildings and hospital wards, and to authenticate medical and social support personnel. There is also an increasing need to identify patients with a high degree of certainty. Finally there is the risk that biometric authentication devices can significantly reveal any health information. All these issues require a careful ethical and political scrutiny.

  15. Approaches to advancescientific understanding of macrosystems ecology

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Levy, Ofir; Ball, Becky; Bond-Lamberty, Benjamin

    Macrosystem ecological studies inherently investigate processes that interact across multiple spatial and temporal scales, requiring intensive sampling and massive amounts of data from diverse sources to incorporate complex cross-scale and hierarchical interactions. Inherent challenges associated with these characteristics include high computational demands, data standardization and assimilation, identification of important processes and scales without prior knowledge, and the need for large, cross-disciplinary research teams that conduct long-term studies. Therefore, macrosystem ecology studies must utilize a unique set of approaches that are capable of encompassing these methodological characteristics and associated challenges. Several case studies demonstrate innovative methods used in current macrosystem ecologymore » studies.« less

  16. Jimena: efficient computing and system state identification for genetic regulatory networks.

    PubMed

    Karl, Stefan; Dandekar, Thomas

    2013-10-11

    Boolean networks capture switching behavior of many naturally occurring regulatory networks. For semi-quantitative modeling, interpolation between ON and OFF states is necessary. The high degree polynomial interpolation of Boolean genetic regulatory networks (GRNs) in cellular processes such as apoptosis or proliferation allows for the modeling of a wider range of node interactions than continuous activator-inhibitor models, but suffers from scaling problems for networks which contain nodes with more than ~10 inputs. Many GRNs from literature or new gene expression experiments exceed those limitations and a new approach was developed. (i) As a part of our new GRN simulation framework Jimena we introduce and setup Boolean-tree-based data structures; (ii) corresponding algorithms greatly expedite the calculation of the polynomial interpolation in almost all cases, thereby expanding the range of networks which can be simulated by this model in reasonable time. (iii) Stable states for discrete models are efficiently counted and identified using binary decision diagrams. As application example, we show how system states can now be sampled efficiently in small up to large scale hormone disease networks (Arabidopsis thaliana development and immunity, pathogen Pseudomonas syringae and modulation by cytokinins and plant hormones). Jimena simulates currently available GRNs about 10-100 times faster than the previous implementation of the polynomial interpolation model and even greater gains are achieved for large scale-free networks. This speed-up also facilitates a much more thorough sampling of continuous state spaces which may lead to the identification of new stable states. Mutants of large networks can be constructed and analyzed very quickly enabling new insights into network robustness and behavior.

  17. Electro-thermal battery model identification for automotive applications

    NASA Astrophysics Data System (ADS)

    Hu, Y.; Yurkovich, S.; Guezennec, Y.; Yurkovich, B. J.

    This paper describes a model identification procedure for identifying an electro-thermal model of lithium ion batteries used in automotive applications. The dynamic model structure adopted is based on an equivalent circuit model whose parameters are scheduled on the state-of-charge, temperature, and current direction. Linear spline functions are used as the functional form for the parametric dependence. The model identified in this way is valid inside a large range of temperatures and state-of-charge, so that the resulting model can be used for automotive applications such as on-board estimation of the state-of-charge and state-of-health. The model coefficients are identified using a multiple step genetic algorithm based optimization procedure designed for large scale optimization problems. The validity of the procedure is demonstrated experimentally for an A123 lithium ion iron-phosphate battery.

  18. Mentors Offering Maternal Support Reduces Prenatal, Pregnancy-Specific Anxiety in a Sample of Military Women.

    PubMed

    Weis, Karen L; Lederman, Regina P; Walker, Katherine C; Chan, Wenyaw

    To determine the efficacy of the Mentors Offering Maternal Support (MOMS) program to reduce pregnancy-specific anxiety and depression and build self-esteem and resilience in military women. Randomized controlled trial with repeated measures. Large military community in Texas. Pregnant women (N = 246) in a military sample defined as active duty or spouse of military personnel. Participants were randomized in the first trimester to the MOMS program or normal prenatal care. Participants attended eight 1-hour sessions every other week during the first, second, and third trimesters of pregnancy. Pregnancy-specific anxiety, depression, self-esteem, and resilience were measured in each trimester. Linear mixed models were used to compare the two-group difference in slope for prenatal anxiety, depression, self-esteem, and resilience. The Prenatal Self-Evaluation Questionnaire was used to measure perinatal anxiety. Rates of prenatal anxiety on the Identification With a Motherhood Role (p = .049) scale and the Preparation for Labor (p = .017) scale were significantly reduced for participants in MOMS. Nulliparous participants showed significantly lower anxiety on the Acceptance of Pregnancy scale and significantly greater anxiety on the Preparation for Labor scale. Single participants had significantly greater anxiety on the Well-Being of Self and Baby in Labor scale, and participants with deployed husbands had significantly greater anxiety on the Identification With a Motherhood Role scale. Participation in the MOMS program reduced pregnancy-specific prenatal anxiety for the dimensions of Identification With a Motherhood Role and Preparation for Labor. Both dimensions of anxiety were previously found to be significantly associated with preterm birth and low birth weight. Military leaders have recognized the urgent need to support military families. Copyright © 2017 AWHONN, the Association of Women's Health, Obstetric and Neonatal Nurses. Published by Elsevier Inc. All rights reserved.

  19. Large-Scale Biomonitoring of Remote and Threatened Ecosystems via High-Throughput Sequencing

    PubMed Central

    Gibson, Joel F.; Shokralla, Shadi; Curry, Colin; Baird, Donald J.; Monk, Wendy A.; King, Ian; Hajibabaei, Mehrdad

    2015-01-01

    Biodiversity metrics are critical for assessment and monitoring of ecosystems threatened by anthropogenic stressors. Existing sorting and identification methods are too expensive and labour-intensive to be scaled up to meet management needs. Alternately, a high-throughput DNA sequencing approach could be used to determine biodiversity metrics from bulk environmental samples collected as part of a large-scale biomonitoring program. Here we show that both morphological and DNA sequence-based analyses are suitable for recovery of individual taxonomic richness, estimation of proportional abundance, and calculation of biodiversity metrics using a set of 24 benthic samples collected in the Peace-Athabasca Delta region of Canada. The high-throughput sequencing approach was able to recover all metrics with a higher degree of taxonomic resolution than morphological analysis. The reduced cost and increased capacity of DNA sequence-based approaches will finally allow environmental monitoring programs to operate at the geographical and temporal scale required by industrial and regulatory end-users. PMID:26488407

  20. Assembly of 500,000 inter-specific catfish expressed sequence tags and large scale gene-associated marker development for whole genome association studies

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Catfish Genome Consortium; Wang, Shaolin; Peatman, Eric

    2010-03-23

    Background-Through the Community Sequencing Program, a catfish EST sequencing project was carried out through a collaboration between the catfish research community and the Department of Energy's Joint Genome Institute. Prior to this project, only a limited EST resource from catfish was available for the purpose of SNP identification. Results-A total of 438,321 quality ESTs were generated from 8 channel catfish (Ictalurus punctatus) and 4 blue catfish (Ictalurus furcatus) libraries, bringing the number of catfish ESTs to nearly 500,000. Assembly of all catfish ESTs resulted in 45,306 contigs and 66,272 singletons. Over 35percent of the unique sequences had significant similarities tomore » known genes, allowing the identification of 14,776 unique genes in catfish. Over 300,000 putative SNPs have been identified, of which approximately 48,000 are high-quality SNPs identified from contigs with at least four sequences and the minor allele presence of at least two sequences in the contig. The EST resource should be valuable for identification of microsatellites, genome annotation, large-scale expression analysis, and comparative genome analysis. Conclusions-This project generated a large EST resource for catfish that captured the majority of the catfish transcriptome. The parallel analysis of ESTs from two closely related Ictalurid catfishes should also provide powerful means for the evaluation of ancient and recent gene duplications, and for the development of high-density microarrays in catfish. The inter- and intra-specific SNPs identified from all catfish EST dataset assembly will greatly benefit the catfish introgression breeding program and whole genome association studies.« less

  1. JHelioviewer: Open-Source Software for Discovery and Image Access in the Petabyte Age (Invited)

    NASA Astrophysics Data System (ADS)

    Mueller, D.; Dimitoglou, G.; Langenberg, M.; Pagel, S.; Dau, A.; Nuhn, M.; Garcia Ortiz, J. P.; Dietert, H.; Schmidt, L.; Hughitt, V. K.; Ireland, J.; Fleck, B.

    2010-12-01

    The unprecedented torrent of data returned by the Solar Dynamics Observatory is both a blessing and a barrier: a blessing for making available data with significantly higher spatial and temporal resolution, but a barrier for scientists to access, browse and analyze them. With such staggering data volume, the data is bound to be accessible only from a few repositories and users will have to deal with data sets effectively immobile and practically difficult to download. From a scientist's perspective this poses three challenges: accessing, browsing and finding interesting data while avoiding the proverbial search for a needle in a haystack. To address these challenges, we have developed JHelioviewer, an open-source visualization software that lets users browse large data volumes both as still images and movies. We did so by deploying an efficient image encoding, storage, and dissemination solution using the JPEG 2000 standard. This solution enables users to access remote images at different resolution levels as a single data stream. Users can view, manipulate, pan, zoom, and overlay JPEG 2000 compressed data quickly, without severe network bandwidth penalties. Besides viewing data, the browser provides third-party metadata and event catalog integration to quickly locate data of interest, as well as an interface to the Virtual Solar Observatory to download science-quality data. As part of the Helioviewer Project, JHelioviewer offers intuitive ways to browse large amounts of heterogeneous data remotely and provides an extensible and customizable open-source platform for the scientific community.

  2. A Rapid Identification Method for Calamine Using Near-Infrared Spectroscopy Based on Multi-Reference Correlation Coefficient Method and Back Propagation Artificial Neural Network.

    PubMed

    Sun, Yangbo; Chen, Long; Huang, Bisheng; Chen, Keli

    2017-07-01

    As a mineral, the traditional Chinese medicine calamine has a similar shape to many other minerals. Investigations of commercially available calamine samples have shown that there are many fake and inferior calamine goods sold on the market. The conventional identification method for calamine is complicated, therefore as a result of the large scale of calamine samples, a rapid identification method is needed. To establish a qualitative model using near-infrared (NIR) spectroscopy for rapid identification of various calamine samples, large quantities of calamine samples including crude products, counterfeits and processed products were collected and correctly identified using the physicochemical and powder X-ray diffraction method. The NIR spectroscopy method was used to analyze these samples by combining the multi-reference correlation coefficient (MRCC) method and the error back propagation artificial neural network algorithm (BP-ANN), so as to realize the qualitative identification of calamine samples. The accuracy rate of the model based on NIR and MRCC methods was 85%; in addition, the model, which took comprehensive multiple factors into consideration, can be used to identify crude calamine products, its counterfeits and processed products. Furthermore, by in-putting the correlation coefficients of multiple references as the spectral feature data of samples into BP-ANN, a BP-ANN model of qualitative identification was established, of which the accuracy rate was increased to 95%. The MRCC method can be used as a NIR-based method in the process of BP-ANN modeling.

  3. Evaluating the implementation of a national disclosure policy for large-scale adverse events in an integrated health care system: identification of gaps and successes.

    PubMed

    Maguire, Elizabeth M; Bokhour, Barbara G; Wagner, Todd H; Asch, Steven M; Gifford, Allen L; Gallagher, Thomas H; Durfee, Janet M; Martinello, Richard A; Elwy, A Rani

    2016-11-11

    Many healthcare organizations have developed disclosure policies for large-scale adverse events, including the Veterans Health Administration (VA). This study evaluated VA's national large-scale disclosure policy and identifies gaps and successes in its implementation. Semi-structured qualitative interviews were conducted with leaders, hospital employees, and patients at nine sites to elicit their perceptions of recent large-scale adverse events notifications and the national disclosure policy. Data were coded using the constructs of the Consolidated Framework for Implementation Research (CFIR). We conducted 97 interviews. Insights included how to handle the communication of large-scale disclosures through multiple levels of a large healthcare organization and manage ongoing communications about the event with employees. Of the 5 CFIR constructs and 26 sub-constructs assessed, seven were prominent in interviews. Leaders and employees specifically mentioned key problem areas involving 1) networks and communications during disclosure, 2) organizational culture, 3) engagement of external change agents during disclosure, and 4) a need for reflecting on and evaluating the policy implementation and disclosure itself. Patients shared 5) preferences for personal outreach by phone in place of the current use of certified letters. All interviewees discussed 6) issues with execution and 7) costs of the disclosure. CFIR analysis reveals key problem areas that need to be addresses during disclosure, including: timely communication patterns throughout the organization, establishing a supportive culture prior to implementation, using patient-approved, effective communications strategies during disclosures; providing follow-up support for employees and patients, and sharing lessons learned.

  4. Identification of Potential Sources of Mercury (Hg) in Farmland Soil Using a Decision Tree Method in China.

    PubMed

    Zhong, Taiyang; Chen, Dongmei; Zhang, Xiuying

    2016-11-09

    Identification of the sources of soil mercury (Hg) on the provincial scale is helpful for enacting effective policies to prevent further contamination and take reclamation measurements. The natural and anthropogenic sources and their contributions of Hg in Chinese farmland soil were identified based on a decision tree method. The results showed that the concentrations of Hg in parent materials were most strongly associated with the general spatial distribution pattern of Hg concentration on a provincial scale. The decision tree analysis gained an 89.70% total accuracy in simulating the influence of human activities on the additions of Hg in farmland soil. Human activities-for example, the production of coke, application of fertilizers, discharge of wastewater, discharge of solid waste, and the production of non-ferrous metals-were the main external sources of a large amount of Hg in the farmland soil.

  5. Identification of Potential Sources of Mercury (Hg) in Farmland Soil Using a Decision Tree Method in China

    PubMed Central

    Zhong, Taiyang; Chen, Dongmei; Zhang, Xiuying

    2016-01-01

    Identification of the sources of soil mercury (Hg) on the provincial scale is helpful for enacting effective policies to prevent further contamination and take reclamation measurements. The natural and anthropogenic sources and their contributions of Hg in Chinese farmland soil were identified based on a decision tree method. The results showed that the concentrations of Hg in parent materials were most strongly associated with the general spatial distribution pattern of Hg concentration on a provincial scale. The decision tree analysis gained an 89.70% total accuracy in simulating the influence of human activities on the additions of Hg in farmland soil. Human activities—for example, the production of coke, application of fertilizers, discharge of wastewater, discharge of solid waste, and the production of non-ferrous metals—were the main external sources of a large amount of Hg in the farmland soil. PMID:27834884

  6. Uncovering Implicit Assumptions: A Large-Scale Study on Students' Mental Models of Diffusion

    ERIC Educational Resources Information Center

    Stains, Marilyne; Sevian, Hannah

    2015-01-01

    Students' mental models of diffusion in a gas phase solution were studied through the use of the Structure and Motion of Matter (SAMM) survey. This survey permits identification of categories of ways students think about the structure of the gaseous solute and solvent, the origin of motion of gas particles, and trajectories of solute particles in…

  7. From drug to protein: using yeast genetics for high-throughput target discovery.

    PubMed

    Armour, Christopher D; Lum, Pek Yee

    2005-02-01

    The budding yeast Saccharomyces cerevisiae has long been an effective eukaryotic model system for understanding basic cellular processes. The genetic tractability and ease of manipulation in the laboratory make yeast well suited for large-scale chemical and genetic screens. Several recent studies describing the use of yeast genetics for high-throughput drug target identification are discussed in this review.

  8. Competence Assessment of Students with Special Educational Needs--Identification of Appropriate Testing Accommodations

    ERIC Educational Resources Information Center

    Südkamp, Anna; Pohl, Steffi; Weinert, Sabine

    2015-01-01

    Including students with special educational needs in learning (SEN-L) is a challenge for large-scale assessments. In order to draw inferences with respect to students with SEN-L and to compare their scores to students in general education, one needs to assure that the measurement model is reliable and that the same construct is measured for…

  9. MRM-DIFF: data processing strategy for differential analysis in large scale MRM-based lipidomics studies.

    PubMed

    Tsugawa, Hiroshi; Ohta, Erika; Izumi, Yoshihiro; Ogiwara, Atsushi; Yukihira, Daichi; Bamba, Takeshi; Fukusaki, Eiichiro; Arita, Masanori

    2014-01-01

    Based on theoretically calculated comprehensive lipid libraries, in lipidomics as many as 1000 multiple reaction monitoring (MRM) transitions can be monitored for each single run. On the other hand, lipid analysis from each MRM chromatogram requires tremendous manual efforts to identify and quantify lipid species. Isotopic peaks differing by up to a few atomic masses further complicate analysis. To accelerate the identification and quantification process we developed novel software, MRM-DIFF, for the differential analysis of large-scale MRM assays. It supports a correlation optimized warping (COW) algorithm to align MRM chromatograms and utilizes quality control (QC) sample datasets to automatically adjust the alignment parameters. Moreover, user-defined reference libraries that include the molecular formula, retention time, and MRM transition can be used to identify target lipids and to correct peak abundances by considering isotopic peaks. Here, we demonstrate the software pipeline and introduce key points for MRM-based lipidomics research to reduce the mis-identification and overestimation of lipid profiles. The MRM-DIFF program, example data set and the tutorials are downloadable at the "Standalone software" section of the PRIMe (Platform for RIKEN Metabolomics, http://prime.psc.riken.jp/) database website.

  10. MRM-DIFF: data processing strategy for differential analysis in large scale MRM-based lipidomics studies

    PubMed Central

    Tsugawa, Hiroshi; Ohta, Erika; Izumi, Yoshihiro; Ogiwara, Atsushi; Yukihira, Daichi; Bamba, Takeshi; Fukusaki, Eiichiro; Arita, Masanori

    2015-01-01

    Based on theoretically calculated comprehensive lipid libraries, in lipidomics as many as 1000 multiple reaction monitoring (MRM) transitions can be monitored for each single run. On the other hand, lipid analysis from each MRM chromatogram requires tremendous manual efforts to identify and quantify lipid species. Isotopic peaks differing by up to a few atomic masses further complicate analysis. To accelerate the identification and quantification process we developed novel software, MRM-DIFF, for the differential analysis of large-scale MRM assays. It supports a correlation optimized warping (COW) algorithm to align MRM chromatograms and utilizes quality control (QC) sample datasets to automatically adjust the alignment parameters. Moreover, user-defined reference libraries that include the molecular formula, retention time, and MRM transition can be used to identify target lipids and to correct peak abundances by considering isotopic peaks. Here, we demonstrate the software pipeline and introduce key points for MRM-based lipidomics research to reduce the mis-identification and overestimation of lipid profiles. The MRM-DIFF program, example data set and the tutorials are downloadable at the “Standalone software” section of the PRIMe (Platform for RIKEN Metabolomics, http://prime.psc.riken.jp/) database website. PMID:25688256

  11. New Genes and New Insights from Old Genes: Update on Alzheimer Disease

    PubMed Central

    Ringman, John M.; Coppola, Giovanni

    2013-01-01

    Purpose of Review: This article discusses the current status of knowledge regarding the genetic basis of Alzheimer disease (AD) with a focus on clinically relevant aspects. Recent Findings: The genetic architecture of AD is complex, as it includes multiple susceptibility genes and likely nongenetic factors. Rare but highly penetrant autosomal dominant mutations explain a small minority of the cases but have allowed tremendous advances in understanding disease pathogenesis. The identification of a strong genetic risk factor, APOE, reshaped the field and introduced the notion of genetic risk for AD. More recently, large-scale genome-wide association studies are adding to the picture a number of common variants with very small effect sizes. Large-scale resequencing studies are expected to identify additional risk factors, including rare susceptibility variants and structural variation. Summary: Genetic assessment is currently of limited utility in clinical practice because of the low frequency (Mendelian mutations) or small effect size (common risk factors) of the currently known susceptibility genes. However, genetic studies are identifying with confidence a number of novel risk genes, and this will further our understanding of disease biology and possibly the identification of therapeutic targets. PMID:23558482

  12. APRICOT: an integrated computational pipeline for the sequence-based identification and characterization of RNA-binding proteins.

    PubMed

    Sharan, Malvika; Förstner, Konrad U; Eulalio, Ana; Vogel, Jörg

    2017-06-20

    RNA-binding proteins (RBPs) have been established as core components of several post-transcriptional gene regulation mechanisms. Experimental techniques such as cross-linking and co-immunoprecipitation have enabled the identification of RBPs, RNA-binding domains (RBDs) and their regulatory roles in the eukaryotic species such as human and yeast in large-scale. In contrast, our knowledge of the number and potential diversity of RBPs in bacteria is poorer due to the technical challenges associated with the existing global screening approaches. We introduce APRICOT, a computational pipeline for the sequence-based identification and characterization of proteins using RBDs known from experimental studies. The pipeline identifies functional motifs in protein sequences using position-specific scoring matrices and Hidden Markov Models of the functional domains and statistically scores them based on a series of sequence-based features. Subsequently, APRICOT identifies putative RBPs and characterizes them by several biological properties. Here we demonstrate the application and adaptability of the pipeline on large-scale protein sets, including the bacterial proteome of Escherichia coli. APRICOT showed better performance on various datasets compared to other existing tools for the sequence-based prediction of RBPs by achieving an average sensitivity and specificity of 0.90 and 0.91 respectively. The command-line tool and its documentation are available at https://pypi.python.org/pypi/bio-apricot. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  13. Caught you: threats to confidentiality due to the public release of large-scale genetic data sets

    PubMed Central

    2010-01-01

    Background Large-scale genetic data sets are frequently shared with other research groups and even released on the Internet to allow for secondary analysis. Study participants are usually not informed about such data sharing because data sets are assumed to be anonymous after stripping off personal identifiers. Discussion The assumption of anonymity of genetic data sets, however, is tenuous because genetic data are intrinsically self-identifying. Two types of re-identification are possible: the "Netflix" type and the "profiling" type. The "Netflix" type needs another small genetic data set, usually with less than 100 SNPs but including a personal identifier. This second data set might originate from another clinical examination, a study of leftover samples or forensic testing. When merged to the primary, unidentified set it will re-identify all samples of that individual. Even with no second data set at hand, a "profiling" strategy can be developed to extract as much information as possible from a sample collection. Starting with the identification of ethnic subgroups along with predictions of body characteristics and diseases, the asthma kids case as a real-life example is used to illustrate that approach. Summary Depending on the degree of supplemental information, there is a good chance that at least a few individuals can be identified from an anonymized data set. Any re-identification, however, may potentially harm study participants because it will release individual genetic disease risks to the public. PMID:21190545

  14. Caught you: threats to confidentiality due to the public release of large-scale genetic data sets.

    PubMed

    Wjst, Matthias

    2010-12-29

    Large-scale genetic data sets are frequently shared with other research groups and even released on the Internet to allow for secondary analysis. Study participants are usually not informed about such data sharing because data sets are assumed to be anonymous after stripping off personal identifiers. The assumption of anonymity of genetic data sets, however, is tenuous because genetic data are intrinsically self-identifying. Two types of re-identification are possible: the "Netflix" type and the "profiling" type. The "Netflix" type needs another small genetic data set, usually with less than 100 SNPs but including a personal identifier. This second data set might originate from another clinical examination, a study of leftover samples or forensic testing. When merged to the primary, unidentified set it will re-identify all samples of that individual. Even with no second data set at hand, a "profiling" strategy can be developed to extract as much information as possible from a sample collection. Starting with the identification of ethnic subgroups along with predictions of body characteristics and diseases, the asthma kids case as a real-life example is used to illustrate that approach. Depending on the degree of supplemental information, there is a good chance that at least a few individuals can be identified from an anonymized data set. Any re-identification, however, may potentially harm study participants because it will release individual genetic disease risks to the public.

  15. An Optimization Code for Nonlinear Transient Problems of a Large Scale Multidisciplinary Mathematical Model

    NASA Astrophysics Data System (ADS)

    Takasaki, Koichi

    This paper presents a program for the multidisciplinary optimization and identification problem of the nonlinear model of large aerospace vehicle structures. The program constructs the global matrix of the dynamic system in the time direction by the p-version finite element method (pFEM), and the basic matrix for each pFEM node in the time direction is described by a sparse matrix similarly to the static finite element problem. The algorithm used by the program does not require the Hessian matrix of the objective function and so has low memory requirements. It also has a relatively low computational cost, and is suited to parallel computation. The program was integrated as a solver module of the multidisciplinary analysis system CUMuLOUS (Computational Utility for Multidisciplinary Large scale Optimization of Undense System) which is under development by the Aerospace Research and Development Directorate (ARD) of the Japan Aerospace Exploration Agency (JAXA).

  16. Large-scale recording of neuronal ensembles.

    PubMed

    Buzsáki, György

    2004-05-01

    How does the brain orchestrate perceptions, thoughts and actions from the spiking activity of its neurons? Early single-neuron recording research treated spike pattern variability as noise that needed to be averaged out to reveal the brain's representation of invariant input. Another view is that variability of spikes is centrally coordinated and that this brain-generated ensemble pattern in cortical structures is itself a potential source of cognition. Large-scale recordings from neuronal ensembles now offer the opportunity to test these competing theoretical frameworks. Currently, wire and micro-machined silicon electrode arrays can record from large numbers of neurons and monitor local neural circuits at work. Achieving the full potential of massively parallel neuronal recordings, however, will require further development of the neuron-electrode interface, automated and efficient spike-sorting algorithms for effective isolation and identification of single neurons, and new mathematical insights for the analysis of network properties.

  17. Multi-scale Material Parameter Identification Using LS-DYNA® and LS-OPT®

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Stander, Nielen; Basudhar, Anirban; Basu, Ushnish

    2015-09-14

    Ever-tightening regulations on fuel economy, and the likely future regulation of carbon emissions, demand persistent innovation in vehicle design to reduce vehicle mass. Classical methods for computational mass reduction include sizing, shape and topology optimization. One of the few remaining options for weight reduction can be found in materials engineering and material design optimization. Apart from considering different types of materials, by adding material diversity and composite materials, an appealing option in automotive design is to engineer steel alloys for the purpose of reducing plate thickness while retaining sufficient strength and ductility required for durability and safety. A project tomore » develop computational material models for advanced high strength steel is currently being executed under the auspices of the United States Automotive Materials Partnership (USAMP) funded by the US Department of Energy. Under this program, new Third Generation Advanced High Strength Steel (i.e., 3GAHSS) are being designed, tested and integrated with the remaining design variables of a benchmark vehicle Finite Element model. The objectives of the project are to integrate atomistic, microstructural, forming and performance models to create an integrated computational materials engineering (ICME) toolkit for 3GAHSS. The mechanical properties of Advanced High Strength Steels (AHSS) are controlled by many factors, including phase composition and distribution in the overall microstructure, volume fraction, size and morphology of phase constituents as well as stability of the metastable retained austenite phase. The complex phase transformation and deformation mechanisms in these steels make the well-established traditional techniques obsolete, and a multi-scale microstructure-based modeling approach following the ICME [0]strategy was therefore chosen in this project. Multi-scale modeling as a major area of research and development is an outgrowth of the Comprehensive Test Ban Treaty of 1996 which banned surface testing of nuclear devices [1]. This had the effect that experimental work was reduced from large scale tests to multiscale experiments to provide material models with validation at different length scales. In the subsequent years industry realized that multi-scale modeling and simulation-based design were transferable to the design optimization of any structural system. Horstemeyer [1] lists a number of advantages of the use of multiscale modeling. Among these are: the reduction of product development time by alleviating costly trial-and-error iterations as well as the reduction of product costs through innovations in material, product and process designs. Multi-scale modeling can reduce the number of costly large scale experiments and can increase product quality by providing more accurate predictions. Research tends to be focussed on each particular length scale, which enhances accuracy in the long term. This paper serves as an introduction to the LS-OPT and LS-DYNA methodology for multi-scale modeling. It mainly focuses on an approach to integrate material identification using material models of different length scales. As an example, a multi-scale material identification strategy, consisting of a Crystal Plasticity (CP) material model and a homogenized State Variable (SV) model, is discussed and the parameter identification of the individual material models of different length scales is demonstrated. The paper concludes with thoughts on integrating the multi-scale methodology into the overall vehicle design.« less

  18. hEIDI: An Intuitive Application Tool To Organize and Treat Large-Scale Proteomics Data.

    PubMed

    Hesse, Anne-Marie; Dupierris, Véronique; Adam, Claire; Court, Magali; Barthe, Damien; Emadali, Anouk; Masselon, Christophe; Ferro, Myriam; Bruley, Christophe

    2016-10-07

    Advances in high-throughput proteomics have led to a rapid increase in the number, size, and complexity of the associated data sets. Managing and extracting reliable information from such large series of data sets require the use of dedicated software organized in a consistent pipeline to reduce, validate, exploit, and ultimately export data. The compilation of multiple mass-spectrometry-based identification and quantification results obtained in the context of a large-scale project represents a real challenge for developers of bioinformatics solutions. In response to this challenge, we developed a dedicated software suite called hEIDI to manage and combine both identifications and semiquantitative data related to multiple LC-MS/MS analyses. This paper describes how, through a user-friendly interface, hEIDI can be used to compile analyses and retrieve lists of nonredundant protein groups. Moreover, hEIDI allows direct comparison of series of analyses, on the basis of protein groups, while ensuring consistent protein inference and also computing spectral counts. hEIDI ensures that validated results are compliant with MIAPE guidelines as all information related to samples and results is stored in appropriate databases. Thanks to the database structure, validated results generated within hEIDI can be easily exported in the PRIDE XML format for subsequent publication. hEIDI can be downloaded from http://biodev.extra.cea.fr/docs/heidi .

  19. Genetic diversity of armored scales (Hemiptera: Diaspididae) and soft scales (Hemiptera: Coccidae) in Chile.

    PubMed

    Amouroux, P; Crochard, D; Germain, J-F; Correa, M; Ampuero, J; Groussier, G; Kreiter, P; Malausa, T; Zaviezo, T

    2017-05-17

    Scale insects (Sternorrhyncha: Coccoidea) are one of the most invasive and agriculturally damaging insect groups. Their management and the development of new control methods are currently jeopardized by the scarcity of identification data, in particular in regions where no large survey coupling morphological and DNA analyses have been performed. In this study, we sampled 116 populations of armored scales (Hemiptera: Diaspididae) and 112 populations of soft scales (Hemiptera: Coccidae) in Chile, over a latitudinal gradient ranging from 18°S to 41°S, on fruit crops, ornamental plants and trees. We sequenced the COI and 28S genes in each population. In total, 19 Diaspididae species and 11 Coccidae species were identified morphologically. From the 63 COI haplotypes and the 54 28S haplotypes uncovered, and using several DNA data analysis methods (Automatic Barcode Gap Discovery, K2P distance, NJ trees), up to 36 genetic clusters were detected. Morphological and DNA data were congruent, except for three species (Aspidiotus nerii, Hemiberlesia rapax and Coccus hesperidum) in which DNA data revealed highly differentiated lineages. More than 50% of the haplotypes obtained had no high-scoring matches with any of the sequences in the GenBank database. This study provides 63 COI and 54 28S barcode sequences for the identification of Coccoidea from Chile.

  20. Test of the CLAS12 RICH large-scale prototype in the direct proximity focusing configuration

    DOE PAGES

    Anefalos Pereira, S.; Baltzell, N.; Barion, L.; ...

    2016-02-11

    A large area ring-imaging Cherenkov detector has been designed to provide clean hadron identification capability in the momentum range from 3 GeV/c up to 8 GeV/c for the CLAS12 experiments at the upgraded 12 GeV continuous electron beam accelerator facility of Jefferson Laboratory. The adopted solution foresees a novel hybrid optics design based on aerogel radiator, composite mirrors and high-packed and high-segmented photon detectors. Cherenkov light will either be imaged directly (forward tracks) or after two mirror reflections (large angle tracks). We report here the results of the tests of a large scale prototype of the RICH detector performed withmore » the hadron beam of the CERN T9 experimental hall for the direct detection configuration. As a result, the tests demonstrated that the proposed design provides the required pion-to-kaon rejection factor of 1:500 in the whole momentum range.« less

  1. A Framework for Spatial Interaction Analysis Based on Large-Scale Mobile Phone Data

    PubMed Central

    Li, Weifeng; Cheng, Xiaoyun; Guo, Gaohua

    2014-01-01

    The overall understanding of spatial interaction and the exact knowledge of its dynamic evolution are required in the urban planning and transportation planning. This study aimed to analyze the spatial interaction based on the large-scale mobile phone data. The newly arisen mass dataset required a new methodology which was compatible with its peculiar characteristics. A three-stage framework was proposed in this paper, including data preprocessing, critical activity identification, and spatial interaction measurement. The proposed framework introduced the frequent pattern mining and measured the spatial interaction by the obtained association. A case study of three communities in Shanghai was carried out as verification of proposed method and demonstration of its practical application. The spatial interaction patterns and the representative features proved the rationality of the proposed framework. PMID:25435865

  2. miRNAFold: a web server for fast miRNA precursor prediction in genomes.

    PubMed

    Tav, Christophe; Tempel, Sébastien; Poligny, Laurent; Tahi, Fariza

    2016-07-08

    Computational methods are required for prediction of non-coding RNAs (ncRNAs), which are involved in many biological processes, especially at post-transcriptional level. Among these ncRNAs, miRNAs have been largely studied and biologists need efficient and fast tools for their identification. In particular, ab initio methods are usually required when predicting novel miRNAs. Here we present a web server dedicated for miRNA precursors identification at a large scale in genomes. It is based on an algorithm called miRNAFold that allows predicting miRNA hairpin structures quickly with high sensitivity. miRNAFold is implemented as a web server with an intuitive and user-friendly interface, as well as a standalone version. The web server is freely available at: http://EvryRNA.ibisc.univ-evry.fr/miRNAFold. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  3. The Deviant Organization and the Bad Apple CEO: Ideology and Accountability in Media Coverage of Corporate Scandals

    ERIC Educational Resources Information Center

    Benediktsson, Michael Owen

    2010-01-01

    What role do the media play in the identification and construction of white-collar crimes? Few studies have examined media coverage of corporate deviance. This study investigates news coverage of six large-scale accounting scandals that broke in 2001 and 2002. Using a variety of empirical methods to analyze the 51 largest U.S. newspapers, the…

  4. GHEP-ISFG collaborative simulated exercise for DVI/MPI: Lessons learned about large-scale profile database comparisons.

    PubMed

    Vullo, Carlos M; Romero, Magdalena; Catelli, Laura; Šakić, Mustafa; Saragoni, Victor G; Jimenez Pleguezuelos, María Jose; Romanini, Carola; Anjos Porto, Maria João; Puente Prieto, Jorge; Bofarull Castro, Alicia; Hernandez, Alexis; Farfán, María José; Prieto, Victoria; Alvarez, David; Penacino, Gustavo; Zabalza, Santiago; Hernández Bolaños, Alejandro; Miguel Manterola, Irati; Prieto, Lourdes; Parsons, Thomas

    2016-03-01

    The GHEP-ISFG Working Group has recognized the importance of assisting DNA laboratories to gain expertise in handling DVI or missing persons identification (MPI) projects which involve the need for large-scale genetic profile comparisons. Eleven laboratories participated in a DNA matching exercise to identify victims from a hypothetical conflict with 193 missing persons. The post mortem database was comprised of 87 skeletal remain profiles from a secondary mass grave displaying a minimal number of 58 individuals with evidence of commingling. The reference database was represented by 286 family reference profiles with diverse pedigrees. The goal of the exercise was to correctly discover re-associations and family matches. The results of direct matching for commingled remains re-associations were correct and fully concordant among all laboratories. However, the kinship analysis for missing persons identifications showed variable results among the participants. There was a group of laboratories with correct, concordant results but nearly half of the others showed discrepant results exhibiting likelihood ratio differences of several degrees of magnitude in some cases. Three main errors were detected: (a) some laboratories did not use the complete reference family genetic data to report the match with the remains, (b) the identity and/or non-identity hypotheses were sometimes wrongly expressed in the likelihood ratio calculations, and (c) many laboratories did not properly evaluate the prior odds for the event. The results suggest that large-scale profile comparisons for DVI or MPI is a challenge for forensic genetics laboratories and the statistical treatment of DNA matching and the Bayesian framework should be better standardized among laboratories. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  5. Leveraging 3D chemical similarity, target and phenotypic data in the identification of drug-protein and drug-adverse effect associations.

    PubMed

    Vilar, Santiago; Hripcsak, George

    2016-01-01

    Drug-target identification is crucial to discover novel applications for existing drugs and provide more insights about mechanisms of biological actions, such as adverse drug effects (ADEs). Computational methods along with the integration of current big data sources provide a useful framework for drug-target and drug-adverse effect discovery. In this article, we propose a method based on the integration of 3D chemical similarity, target and adverse effect data to generate a drug-target-adverse effect predictor along with a simple leveraging system to improve identification of drug-targets and drug-adverse effects. In the first step, we generated a system for multiple drug-target identification based on the application of 3D drug similarity into a large target dataset extracted from the ChEMBL. Next, we developed a target-adverse effect predictor combining targets from ChEMBL with phenotypic information provided by SIDER data source. Both modules were linked to generate a final predictor that establishes hypothesis about new drug-target-adverse effect candidates. Additionally, we showed that leveraging drug-target candidates with phenotypic data is very useful to improve the identification of drug-targets. The integration of phenotypic data into drug-target candidates yielded up to twofold precision improvement. In the opposite direction, leveraging drug-phenotype candidates with target data also yielded a significant enhancement in the performance. The modeling described in the current study is simple and efficient and has applications at large scale in drug repurposing and drug safety through the identification of mechanism of action of biological effects.

  6. Monitoring scale scores over time via quality control charts, model-based approaches, and time series techniques.

    PubMed

    Lee, Yi-Hsuan; von Davier, Alina A

    2013-07-01

    Maintaining a stable score scale over time is critical for all standardized educational assessments. Traditional quality control tools and approaches for assessing scale drift either require special equating designs, or may be too time-consuming to be considered on a regular basis with an operational test that has a short time window between an administration and its score reporting. Thus, the traditional methods are not sufficient to catch unusual testing outcomes in a timely manner. This paper presents a new approach for score monitoring and assessment of scale drift. It involves quality control charts, model-based approaches, and time series techniques to accommodate the following needs of monitoring scale scores: continuous monitoring, adjustment of customary variations, identification of abrupt shifts, and assessment of autocorrelation. Performance of the methodologies is evaluated using manipulated data based on real responses from 71 administrations of a large-scale high-stakes language assessment.

  7. A Census of Large-scale (≥10 PC), Velocity-coherent, Dense Filaments in the Northern Galactic Plane: Automated Identification Using Minimum Spanning Tree

    NASA Astrophysics Data System (ADS)

    Wang, Ke; Testi, Leonardo; Burkert, Andreas; Walmsley, C. Malcolm; Beuther, Henrik; Henning, Thomas

    2016-09-01

    Large-scale gaseous filaments with lengths up to the order of 100 pc are on the upper end of the filamentary hierarchy of the Galactic interstellar medium (ISM). Their association with respect to the Galactic structure and their role in Galactic star formation are of great interest from both an observational and theoretical point of view. Previous “by-eye” searches, combined together, have started to uncover the Galactic distribution of large filaments, yet inherent bias and small sample size limit conclusive statistical results from being drawn. Here, we present (1) a new, automated method for identifying large-scale velocity-coherent dense filaments, and (2) the first statistics and the Galactic distribution of these filaments. We use a customized minimum spanning tree algorithm to identify filaments by connecting voxels in the position-position-velocity space, using the Bolocam Galactic Plane Survey spectroscopic catalog. In the range of 7\\buildrel{\\circ}\\over{.} 5≤slant l≤slant 194^\\circ , we have identified 54 large-scale filaments and derived mass (˜ {10}3{--}{10}5 {M}⊙ ), length (10-276 pc), linear mass density (54-8625 {M}⊙ pc-1), aspect ratio, linearity, velocity gradient, temperature, fragmentation, Galactic location, and orientation angle. The filaments concentrate along major spiral arms. They are widely distributed across the Galactic disk, with 50% located within ±20 pc from the Galactic mid-plane and 27% run in the center of spiral arms. An order of 1% of the molecular ISM is confined in large filaments. Massive star formation is more favorable in large filaments compared to elsewhere. This is the first comprehensive catalog of large filaments that can be useful for a quantitative comparison with spiral structures and numerical simulations.

  8. WISDOM-II: screening against multiple targets implicated in malaria using computational grid infrastructures.

    PubMed

    Kasam, Vinod; Salzemann, Jean; Botha, Marli; Dacosta, Ana; Degliesposti, Gianluca; Isea, Raul; Kim, Doman; Maass, Astrid; Kenyon, Colin; Rastelli, Giulio; Hofmann-Apitius, Martin; Breton, Vincent

    2009-05-01

    Despite continuous efforts of the international community to reduce the impact of malaria on developing countries, no significant progress has been made in the recent years and the discovery of new drugs is more than ever needed. Out of the many proteins involved in the metabolic activities of the Plasmodium parasite, some are promising targets to carry out rational drug discovery. Recent years have witnessed the emergence of grids, which are highly distributed computing infrastructures particularly well fitted for embarrassingly parallel computations like docking. In 2005, a first attempt at using grids for large-scale virtual screening focused on plasmepsins and ended up in the identification of previously unknown scaffolds, which were confirmed in vitro to be active plasmepsin inhibitors. Following this success, a second deployment took place in the fall of 2006 focussing on one well known target, dihydrofolate reductase (DHFR), and on a new promising one, glutathione-S-transferase. In silico drug design, especially vHTS is a widely and well-accepted technology in lead identification and lead optimization. This approach, therefore builds, upon the progress made in computational chemistry to achieve more accurate in silico docking and in information technology to design and operate large scale grid infrastructures. On the computational side, a sustained infrastructure has been developed: docking at large scale, using different strategies in result analysis, storing of the results on the fly into MySQL databases and application of molecular dynamics refinement are MM-PBSA and MM-GBSA rescoring. The modeling results obtained are very promising. Based on the modeling results, In vitro results are underway for all the targets against which screening is performed. The current paper describes the rational drug discovery activity at large scale, especially molecular docking using FlexX software on computational grids in finding hits against three different targets (PfGST, PfDHFR, PvDHFR (wild type and mutant forms) implicated in malaria. Grid-enabled virtual screening approach is proposed to produce focus compound libraries for other biological targets relevant to fight the infectious diseases of the developing world.

  9. Implementation of the Agitated Behavior Scale in the Electronic Health Record.

    PubMed

    Wilson, Helen John; Dasgupta, Kritis; Michael, Kathleen

    The purpose of the study was to implement an Agitated Behavior Scale through an electronic health record and to evaluate the usability of the scale in a brain injury unit at a rehabilitation hospital. A quality improvement project was conducted in the brain injury unit at a large rehabilitation hospital with registered nurses as participants using convenience sampling. The project consisted of three phases and included education, implementation of the scale in the electronic health record, and administration of the survey questionnaire, which utilized the system usability scale. The Agitated Behavior Scale was found to be usable, and there was 92.2% compliance with the use of the electronic Electronic Agitated Behavior Scale. The Agitated Behavior Scale was effectively implemented in the electronic health record and was found to be usable in the assessment of agitation. Utilization of the scale through the electronic health record on a daily basis will allow for an early identification of agitation in patients with traumatic brain injury and enable prompt interventions to manage agitation.

  10. Estimating the reliability of eyewitness identifications from police lineups

    PubMed Central

    Wixted, John T.; Mickes, Laura; Dunn, John C.; Clark, Steven E.; Wells, William

    2016-01-01

    Laboratory-based mock crime studies have often been interpreted to mean that (i) eyewitness confidence in an identification made from a lineup is a weak indicator of accuracy and (ii) sequential lineups are diagnostically superior to traditional simultaneous lineups. Largely as a result, juries are increasingly encouraged to disregard eyewitness confidence, and up to 30% of law enforcement agencies in the United States have adopted the sequential procedure. We conducted a field study of actual eyewitnesses who were assigned to simultaneous or sequential photo lineups in the Houston Police Department over a 1-y period. Identifications were made using a three-point confidence scale, and a signal detection model was used to analyze and interpret the results. Our findings suggest that (i) confidence in an eyewitness identification from a fair lineup is a highly reliable indicator of accuracy and (ii) if there is any difference in diagnostic accuracy between the two lineup formats, it likely favors the simultaneous procedure. PMID:26699467

  11. Estimating the reliability of eyewitness identifications from police lineups.

    PubMed

    Wixted, John T; Mickes, Laura; Dunn, John C; Clark, Steven E; Wells, William

    2016-01-12

    Laboratory-based mock crime studies have often been interpreted to mean that (i) eyewitness confidence in an identification made from a lineup is a weak indicator of accuracy and (ii) sequential lineups are diagnostically superior to traditional simultaneous lineups. Largely as a result, juries are increasingly encouraged to disregard eyewitness confidence, and up to 30% of law enforcement agencies in the United States have adopted the sequential procedure. We conducted a field study of actual eyewitnesses who were assigned to simultaneous or sequential photo lineups in the Houston Police Department over a 1-y period. Identifications were made using a three-point confidence scale, and a signal detection model was used to analyze and interpret the results. Our findings suggest that (i) confidence in an eyewitness identification from a fair lineup is a highly reliable indicator of accuracy and (ii) if there is any difference in diagnostic accuracy between the two lineup formats, it likely favors the simultaneous procedure.

  12. From GenBank to GBIF: Phylogeny-Based Predictive Niche Modeling Tests Accuracy of Taxonomic Identifications in Large Occurrence Data Repositories

    PubMed Central

    Smith, B. Eugene; Johnston, Mark K.; Lücking, Robert

    2016-01-01

    Accuracy of taxonomic identifications is crucial to data quality in online repositories of species occurrence data, such as the Global Biodiversity Information Facility (GBIF), which have accumulated several hundred million records over the past 15 years. These data serve as basis for large scale analyses of macroecological and biogeographic patterns and to document environmental changes over time. However, taxonomic identifications are often unreliable, especially for non-vascular plants and fungi including lichens, which may lack critical revisions of voucher specimens. Due to the scale of the problem, restudy of millions of collections is unrealistic and other strategies are needed. Here we propose to use verified, georeferenced occurrence data of a given species to apply predictive niche modeling that can then be used to evaluate unverified occurrences of that species. Selecting the charismatic lichen fungus, Usnea longissima, as a case study, we used georeferenced occurrence records based on sequenced specimens to model its predicted niche. Our results suggest that the target species is largely restricted to a narrow range of boreal and temperate forest in the Northern Hemisphere and that occurrence records in GBIF from tropical regions and the Southern Hemisphere do not represent this taxon, a prediction tested by comparison with taxonomic revisions of Usnea for these regions. As a novel approach, we employed Principal Component Analysis on the environmental grid data used for predictive modeling to visualize potential ecogeographical barriers for the target species; we found that tropical regions conform a strong barrier, explaining why potential niches in the Southern Hemisphere were not colonized by Usnea longissima and instead by morphologically similar species. This approach is an example of how data from two of the most important biodiversity repositories, GenBank and GBIF, can be effectively combined to remotely address the problem of inaccuracy of taxonomic identifications in occurrence data repositories and to provide a filtering mechanism which can considerably reduce the number of voucher specimens that need critical revision, in this case from 4,672 to about 100. PMID:26967999

  13. From GenBank to GBIF: Phylogeny-Based Predictive Niche Modeling Tests Accuracy of Taxonomic Identifications in Large Occurrence Data Repositories.

    PubMed

    Smith, B Eugene; Johnston, Mark K; Lücking, Robert

    2016-01-01

    Accuracy of taxonomic identifications is crucial to data quality in online repositories of species occurrence data, such as the Global Biodiversity Information Facility (GBIF), which have accumulated several hundred million records over the past 15 years. These data serve as basis for large scale analyses of macroecological and biogeographic patterns and to document environmental changes over time. However, taxonomic identifications are often unreliable, especially for non-vascular plants and fungi including lichens, which may lack critical revisions of voucher specimens. Due to the scale of the problem, restudy of millions of collections is unrealistic and other strategies are needed. Here we propose to use verified, georeferenced occurrence data of a given species to apply predictive niche modeling that can then be used to evaluate unverified occurrences of that species. Selecting the charismatic lichen fungus, Usnea longissima, as a case study, we used georeferenced occurrence records based on sequenced specimens to model its predicted niche. Our results suggest that the target species is largely restricted to a narrow range of boreal and temperate forest in the Northern Hemisphere and that occurrence records in GBIF from tropical regions and the Southern Hemisphere do not represent this taxon, a prediction tested by comparison with taxonomic revisions of Usnea for these regions. As a novel approach, we employed Principal Component Analysis on the environmental grid data used for predictive modeling to visualize potential ecogeographical barriers for the target species; we found that tropical regions conform a strong barrier, explaining why potential niches in the Southern Hemisphere were not colonized by Usnea longissima and instead by morphologically similar species. This approach is an example of how data from two of the most important biodiversity repositories, GenBank and GBIF, can be effectively combined to remotely address the problem of inaccuracy of taxonomic identifications in occurrence data repositories and to provide a filtering mechanism which can considerably reduce the number of voucher specimens that need critical revision, in this case from 4,672 to about 100.

  14. Model and Data Reduction for Control, Identification and Compressed Sensing

    NASA Astrophysics Data System (ADS)

    Kramer, Boris

    This dissertation focuses on problems in design, optimization and control of complex, large-scale dynamical systems from different viewpoints. The goal is to develop new algorithms and methods, that solve real problems more efficiently, together with providing mathematical insight into the success of those methods. There are three main contributions in this dissertation. In Chapter 3, we provide a new method to solve large-scale algebraic Riccati equations, which arise in optimal control, filtering and model reduction. We present a projection based algorithm utilizing proper orthogonal decomposition, which is demonstrated to produce highly accurate solutions at low rank. The method is parallelizable, easy to implement for practitioners, and is a first step towards a matrix free approach to solve AREs. Numerical examples for n ≥ 106 unknowns are presented. In Chapter 4, we develop a system identification method which is motivated by tangential interpolation. This addresses the challenge of fitting linear time invariant systems to input-output responses of complex dynamics, where the number of inputs and outputs is relatively large. The method reduces the computational burden imposed by a full singular value decomposition, by carefully choosing directions on which to project the impulse response prior to assembly of the Hankel matrix. The identification and model reduction step follows from the eigensystem realization algorithm. We present three numerical examples, a mass spring damper system, a heat transfer problem, and a fluid dynamics system. We obtain error bounds and stability results for this method. Chapter 5 deals with control and observation design for parameter dependent dynamical systems. We address this by using local parametric reduced order models, which can be used online. Data available from simulations of the system at various configurations (parameters, boundary conditions) is used to extract a sparse basis to represent the dynamics (via dynamic mode decomposition). Subsequently, a new, compressed sensing based classification algorithm is developed which incorporates the extracted dynamic information into the sensing basis. We show that this augmented classification basis makes the method more robust to noise, and results in superior identification of the correct parameter. Numerical examples consist of a Navier-Stokes, as well as a Boussinesq flow application.

  15. Can rove beetles (Staphylinidae) be excluded in studies focusing on saproxylic beetles in central European beech forests?

    PubMed

    Parmain, G; Bouget, C; Müller, J; Horak, J; Gossner, M M; Lachat, T; Isacsson, G

    2015-02-01

    Monitoring saproxylic beetle diversity, though challenging, can help identifying relevant conservation sites or key drivers of forest biodiversity, and assessing the impact of forestry practices on biodiversity. Unfortunately, monitoring species assemblages is costly, mainly due to the time spent on identification. Excluding families which are rich in specimens and species but are difficult to identify is a frequent procedure used in ecological entomology to reduce the identification cost. The Staphylinidae (rove beetle) family is both one of the most frequently excluded and one of the most species-rich saproxylic beetle families. Using a large-scale beetle and environmental dataset from 238 beech stands across Europe, we evaluated the effects of staphylinid exclusion on results in ecological forest studies. Simplified staphylinid-excluded assemblages were found to be relevant surrogates for whole assemblages. The species richness and composition of saproxylic beetle assemblages both with and without staphylinids responded congruently to landscape, climatic and stand gradients, even when the assemblages included a high proportion of staphylinid species. At both local and regional scales, the species richness as well as the species composition of staphylinid-included and staphylinid-excluded assemblages were highly positively correlated. Ranking of sites according to their biodiversity level, which either included or excluded Staphylinidae in species richness, also gave congruent results. From our results, species assemblages omitting staphylinids can be taken as efficient surrogates for complete assemblages in large scale biodiversity monitoring studies.

  16. Development of a New Marker System for Identification of Spirodela polyrhiza and Landoltia punctata

    PubMed Central

    Feng, Bo; Fang, Yang; Xu, Zhibin; Xiang, Chao; Zhou, Chunhong; Jiang, Fei; Wang, Tao

    2017-01-01

    Lemnaceae (commonly called duckweed) is an aquatic plant ideal for quantitative analysis in plant sciences. Several species of this family represent the smallest and fastest growing flowering plants. Different ecotypes of the same species vary in their biochemical and physiological properties. Thus, selecting of desirable ecotypes of a species is very important. Here, we developed a simple and rapid molecular identification system for Spirodela polyrhiza and Landoltia punctata based on the sequence polymorphism. First, several pairs of primers were designed and three markers were selected as good for identification. After PCR amplification, DNA fragments (the combination of three PCR products) in different duckweeds were detected using capillary electrophoresis. The high-resolution capillary electrophoresis displayed high identity to the sequencing results. The combination of the PCR products containing several DNA fragments highly improved the identification frequency. These results indicate that this method is not only good for interspecies identification but also ideal for intraspecies distinguishing. Meanwhile, 11 haplotypes were found in both the S. polyrhiza and L. punctata ecotypes. The results suggest that this marker system is useful for large-scale identification of duckweed and for the screening of desirable ecotypes to improve the diverse usage in duckweed utilization. PMID:28168191

  17. Carapace surface architecture facilitates camouflage of the decorator crab Tiarinia cornigera.

    PubMed

    Sanka, Immanuel; Suyono, Eko Agus; Rivero-Müller, Adolfo; Alam, Parvez

    2016-09-01

    This paper elucidates the unique setal morphology of the decorator crab Tiarinia cornigera, and further presents evidence to that setal morphology promotes micro-organism nucleation and adhesion. The carapace of this crab is covered by clusters of setae, each comprising a hollow acicular stem that is enveloped by a haystack-like structure. Using computational fluid dynamics, we find that these setae are responsible for manipulating water flow over the carapace surface. Micro-organisms in the sea water, nest in areas of flow stagnation and as a result, nucleate to and biofoul the setae by means of chemical adhesion. Attached micro-organisms secrete extracellular polymeric substances, which we deduce must also provide an additional element of chemical adhesion to mechanically interlocked mesoscopic and macroscopic biomatter. By coupling physical and chemical methods for adhesion, T. cornigera is able to hierarchically decorate its carapace. Our paper brings to light the unique decorator crab carapace morphology of T. cornigera; and furthermore evidences its function in micro-organism nucleation and adhesion. We show how this special carapace morphology directs and guides water flow to form nesting regions of water stagnation where micro-organisms can nucleate and adhere. In the literature, decorator crab carapaces are presumed to be able to mechanically interlock biomatter as camouflage using hook-like setal outgrowths. T. cornigera contrarily exhibits clusters of hay-stack like structures. By encouraging micro-organism adhesion to the carapace setae, T. cornigera is able to effectively attach biomatter using both chemical and physical principles of adhesion. T. cornigera essentially has a super-biofouling carapace surface, for at least micro-organisms. Our work will have an impact on researchers interested in biofouling, adhesion, biomedical and purification filter systems, and in the development of novel biomimetic surfaces with tailored properties. Copyright © 2016 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved.

  18. Solar Eclipse-Induced Changes in the Ionosphere over the Continental US

    NASA Astrophysics Data System (ADS)

    Erickson, P. J.; Zhang, S.; Goncharenko, L. P.; Coster, A. J.; Hysell, D. L.; Sulzer, M. P.; Vierinen, J.

    2017-12-01

    For the first time in 26 years, a total solar eclipse occurred over the continental United States on 21 August 2017, between 16:00-20:00 UT. We report on American solar eclipse observations of the upper atmosphere, conducted by a team led by MIT Haystack Observatory. Efforts measured ionospheric and thermospheric eclipse perturbations. Although eclipse effects have been studied for more than 50 years, recent major sensitivity and resolution advances using radio-based techniques are providing new information on the eclipse ionosphere-thermosphere-mesosphere (ITM) system response. Our study was focused on quantifying eclipse effects on (1) traveling ionospheric disturbances (TIDs) and atmospheric gravity waves (AGWs); (2) spatial ionospheric variations associated with the eclipse; and (3) altitudinal and temporal ionospheric profile variations. We present selected early findings on ITM eclipse response including a dense global network of 6000 GNSS total electron content (TEC) receivers (100 million measurements per day; 1x1 degree spatial grid) and the Millstone Hill and Arecibo incoherent scatter radars. TEC depletions of up to 60% in magnitude were associated with the eclipse umbra and penumbra and consistently trailed the eclipse totality center. TEC enhancements associated with prominent orographic features were observed in the western US due to complex interactions as the lower atmosphere cooled in response to decreasing EUV energy inputs. Strong TIDs in the form of bow waves, stern waves, and a stern wake were observed in TEC data. Altitude-resolved plasma parameter profiles from Millstone Hill saw a nearly 50% decrease in F region electron density in vertical profiles, accompanied by a corresponding 200-250 K decrease in electron temperature. Wide field Millstone Hill radar scans showed similar decreases in electron density to the southwest, maximizing along the line of closest approach to totality. Data is available to the research community through the MIT Haystack Madrigal system. Alongside a summary of observations, we will also present preliminary quantitative comparisons with several ongoing modeling efforts.

  19. Crustal dynamics project session 4 validation and intercomparison experiments 1979-1980 report

    NASA Technical Reports Server (NTRS)

    Liebrecht, P.; Kolenkiewicz, R.; Ryan, J.; Hothem, L.

    1983-01-01

    As part of the Crustal Dynamics Project, an experiment was performed to verify the ability of Satellite Laser Ranging (SLR), Very Long Baseline interferometry (VLBI) and Doppler Satellite Positioning System (Doppler) techniques to estimate the baseline distances between several locations. The Goddard Space Flight Center (GSFC) lasers were in operation at all five sites available to them. The ten baselines involved were analyzed using monthly orbits and various methods of selecting data. The standard deviation of the monthly SLR baseline lengths was at the 7 cm level. The GSFC VLBI (Mark III) data was obtained during three separate experiments. November 1979 at Haystack and Owens Valley, and April and July 1980 at Haystack, Owens Valley, and Fort Davis. Repeatability of the VLBI in determining baseline lengths was calculated to be at the 2 cm level. Jet Propulsion Laboratory (JPL) VLBI (Mark II) data was acquired on the Owens Valley to Goldstone baseline on ten occasions between August 1979 and November 1980. The repeatability of these baseline length determinations was calculated to be at the 5 cm level. National Geodetic Survey (NGS) Doppler data was acquired at all five sites in January 1980. Repeatability of the Doppler determined baseline lengths results were calculated at approximately 30 cm. An intercomparison between baseline distances and associated parameters was made utilizing SLR, VLBI, and Doppler results on all available baselines. The VLBI and SLR length determinations were compared on four baselines with a resultant mean difference of -1 cm and a maximum difference of 12 cm. The SLR and Doppler length determinations were compared on ten baselines with a resultant mean difference of about 30 cm and a maximum difference of about 60 cm. The VLBI and Doppler lengths from seven baselines showed a resultant mean difference of about 30 cm and maximum difference of about 1 meter. The intercomparison of baseline orientation parameters were consistent with past analysis.

  20. Molecules in Space: A Chemistry lab using Radio Astronomy

    NASA Astrophysics Data System (ADS)

    Lekberg, M. J.; Pratap, P.

    2000-12-01

    We present the results of a laboratory exercise developed with the support of the NSF Research Experiences for Teachers program at MIT Haystack Observatory. The exercise takes the students beyond the traditional test tubes of a chemistry laboratory into the interstellar medium, where the same principles that they study about in the classroom are found to hold. It also utilizes the true multi-disciplinary nature of radio astronomy and allows the students to realize how much can be learnt by studying the universe at various wavelengths. The astronomical chemistry laboratory is presented wherein students from Chelmsford High School in Massachusetts operate the 37-m telescope at Haystack Observatory via the internet to observe radio signals from galactic chemicals. The laboratory is designed to be the means by which students witness physical evidence for molecular and orbital shapes by observing the radio emission from rotating dipoles. The laboratory described is a lynch pin activity for an integrated unit that moves from the valance shell electron configurations through molecular and orbital geometry to an understanding that many physical and chemical properties of chemicals are ultimately dependent upon the shape/geometry and consequently, dipole of the molecule. Students are expected to interpret and evaluate the nature of molecular dipoles and account for the diversity of rotational spectra using their conceptual knowledge of bonding orbital theory and their knowledge of the electronic atom. Flexibility in the lab allows students to identify individual chemicals by cross referencing radio emission from the galactic sources they have chosen against a prepared catalogue listing or by choosing to "listen" for specific chemicals at exact frequencies. A teacher resource manual containing information and data on a variety of daytime galactic source and individual chemical flux densities of molecular candidates has been prepared. Collaborative exercises and activities, and associated unit topics have also been developed.

  1. The Psychiatric Genomics Consortium Posttraumatic Stress Disorder Workgroup: Posttraumatic Stress Disorder Enters the Age of Large-Scale Genomic Collaboration

    PubMed Central

    Logue, Mark W; Amstadter, Ananda B; Baker, Dewleen G; Duncan, Laramie; Koenen, Karestan C; Liberzon, Israel; Miller, Mark W; Morey, Rajendra A; Nievergelt, Caroline M; Ressler, Kerry J; Smith, Alicia K; Smoller, Jordan W; Stein, Murray B; Sumner, Jennifer A; Uddin, Monica

    2015-01-01

    The development of posttraumatic stress disorder (PTSD) is influenced by genetic factors. Although there have been some replicated candidates, the identification of risk variants for PTSD has lagged behind genetic research of other psychiatric disorders such as schizophrenia, autism, and bipolar disorder. Psychiatric genetics has moved beyond examination of specific candidate genes in favor of the genome-wide association study (GWAS) strategy of very large numbers of samples, which allows for the discovery of previously unsuspected genes and molecular pathways. The successes of genetic studies of schizophrenia and bipolar disorder have been aided by the formation of a large-scale GWAS consortium: the Psychiatric Genomics Consortium (PGC). In contrast, only a handful of GWAS of PTSD have appeared in the literature to date. Here we describe the formation of a group dedicated to large-scale study of PTSD genetics: the PGC-PTSD. The PGC-PTSD faces challenges related to the contingency on trauma exposure and the large degree of ancestral genetic diversity within and across participating studies. Using the PGC analysis pipeline supplemented by analyses tailored to address these challenges, we anticipate that our first large-scale GWAS of PTSD will comprise over 10 000 cases and 30 000 trauma-exposed controls. Following in the footsteps of our PGC forerunners, this collaboration—of a scope that is unprecedented in the field of traumatic stress—will lead the search for replicable genetic associations and new insights into the biological underpinnings of PTSD. PMID:25904361

  2. Two Scales for the Measurement of Mexican-American Identity.

    ERIC Educational Resources Information Center

    Teske, Raymond, Jr.; Nelson, Bardin H.

    The development of scales to measure Mexican American identification with their population is discussed in this paper. The scales measure (1) identification with the Mexican American population using attitudinal items (Identity Scale) and (2) interaction behavior with the Mexican American population (Interaction Scale). The sample consisted of all…

  3. Identification of a basic helix-loop-helix-type transcription regulator gene in Aspergillus oryzae by systematically deleting large chromosomal segments.

    PubMed

    Jin, Feng Jie; Takahashi, Tadashi; Machida, Masayuki; Koyama, Yasuji

    2009-09-01

    We previously developed two methods (loop-out and replacement-type recombination) for generating large-scale chromosomal deletions that can be applied to more effective chromosomal engineering in Aspergillus oryzae. In this study, the replacement-type method is used to systematically delete large chromosomal DNA segments to identify essential and nonessential regions in chromosome 7 (2.93 Mb), which is the smallest A. oryzae chromosome and contains a large number of nonsyntenic blocks. We constructed 12 mutants harboring deletions that spanned 16- to 150-kb segments of chromosome 7 and scored phenotypic changes in the resulting mutants. Among the deletion mutants, strains designated Delta5 and Delta7 displayed clear phenotypic changes involving growth and conidiation. In particular, the Delta5 mutant exhibited vigorous growth and conidiation, potentially beneficial characteristics for certain industrial applications. Further deletion analysis allowed identification of the AO090011000215 gene as the gene responsible for the Delta5 mutant phenotype. The AO090011000215 gene was predicted to encode a helix-loop-helix binding protein belonging to the bHLH family of transcription factors. These results illustrate the potential of the approach for identifying novel functional genes.

  4. VHSIC Electronics and the Cost of Air Force Avionics in the 1990s

    DTIC Science & Technology

    1990-11-01

    circuit. LRM Line replaceable module. LRU Line replaceable unit. LSI Large-scale integration. LSTTL Tow-power Schottky Transitor -to-Transistor Logic...displays, communications/navigation/identification, electronic combat equipment, dispensers, and computers. These CERs, which statistically relate the...some of the reliability numbers, and adding the F-15 and F-16 to obtain the data sample shown in Table 6. Both suite costs and reliability statistics

  5. PATIKA: an integrated visual environment for collaborative construction and analysis of cellular pathways.

    PubMed

    Demir, E; Babur, O; Dogrusoz, U; Gursoy, A; Nisanci, G; Cetin-Atalay, R; Ozturk, M

    2002-07-01

    Availability of the sequences of entire genomes shifts the scientific curiosity towards the identification of function of the genomes in large scale as in genome studies. In the near future, data produced about cellular processes at molecular level will accumulate with an accelerating rate as a result of proteomics studies. In this regard, it is essential to develop tools for storing, integrating, accessing, and analyzing this data effectively. We define an ontology for a comprehensive representation of cellular events. The ontology presented here enables integration of fragmented or incomplete pathway information and supports manipulation and incorporation of the stored data, as well as multiple levels of abstraction. Based on this ontology, we present the architecture of an integrated environment named Patika (Pathway Analysis Tool for Integration and Knowledge Acquisition). Patika is composed of a server-side, scalable, object-oriented database and client-side editors to provide an integrated, multi-user environment for visualizing and manipulating network of cellular events. This tool features automated pathway layout, functional computation support, advanced querying and a user-friendly graphical interface. We expect that Patika will be a valuable tool for rapid knowledge acquisition, microarray generated large-scale data interpretation, disease gene identification, and drug development. A prototype of Patika is available upon request from the authors.

  6. The opportunities and challenges of large-scale molecular approaches to songbird neurobiology

    PubMed Central

    Mello, C.V.; Clayton, D.F.

    2014-01-01

    High-through put methods for analyzing genome structure and function are having a large impact in song-bird neurobiology. Methods include genome sequencing and annotation, comparative genomics, DNA microarrays and transcriptomics, and the development of a brain atlas of gene expression. Key emerging findings include the identification of complex transcriptional programs active during singing, the robust brain expression of non-coding RNAs, evidence of profound variations in gene expression across brain regions, and the identification of molecular specializations within song production and learning circuits. Current challenges include the statistical analysis of large datasets, effective genome curations, the efficient localization of gene expression changes to specific neuronal circuits and cells, and the dissection of behavioral and environmental factors that influence brain gene expression. The field requires efficient methods for comparisons with organisms like chicken, which offer important anatomical, functional and behavioral contrasts. As sequencing costs plummet, opportunities emerge for comparative approaches that may help reveal evolutionary transitions contributing to vocal learning, social behavior and other properties that make songbirds such compelling research subjects. PMID:25280907

  7. Continental-scale patterns of canopy tree composition and function across Amazonia.

    PubMed

    ter Steege, Hans; Pitman, Nigel C A; Phillips, Oliver L; Chave, Jerome; Sabatier, Daniel; Duque, Alvaro; Molino, Jean-François; Prévost, Marie-Françoise; Spichiger, Rodolphe; Castellanos, Hernán; von Hildebrand, Patricio; Vásquez, Rodolfo

    2006-09-28

    The world's greatest terrestrial stores of biodiversity and carbon are found in the forests of northern South America, where large-scale biogeographic patterns and processes have recently begun to be described. Seven of the nine countries with territory in the Amazon basin and the Guiana shield have carried out large-scale forest inventories, but such massive data sets have been little exploited by tropical plant ecologists. Although forest inventories often lack the species-level identifications favoured by tropical plant ecologists, their consistency of measurement and vast spatial coverage make them ideally suited for numerical analyses at large scales, and a valuable resource to describe the still poorly understood spatial variation of biomass, diversity, community composition and forest functioning across the South American tropics. Here we show, by using the seven forest inventories complemented with trait and inventory data collected elsewhere, two dominant gradients in tree composition and function across the Amazon, one paralleling a major gradient in soil fertility and the other paralleling a gradient in dry season length. The data set also indicates that the dominance of Fabaceae in the Guiana shield is not necessarily the result of root adaptations to poor soils (nodulation or ectomycorrhizal associations) but perhaps also the result of their remarkably high seed mass there as a potential adaptation to low rates of disturbance.

  8. Continental-scale patterns of canopy tree composition and function across Amazonia

    NASA Astrophysics Data System (ADS)

    Ter Steege, Hans; Pitman, Nigel C. A.; Phillips, Oliver L.; Chave, Jerome; Sabatier, Daniel; Duque, Alvaro; Molino, Jean-François; Prévost, Marie-Françoise; Spichiger, Rodolphe; Castellanos, Hernán; von Hildebrand, Patricio; Vásquez, Rodolfo

    2006-09-01

    The world's greatest terrestrial stores of biodiversity and carbon are found in the forests of northern South America, where large-scale biogeographic patterns and processes have recently begun to be described. Seven of the nine countries with territory in the Amazon basin and the Guiana shield have carried out large-scale forest inventories, but such massive data sets have been little exploited by tropical plant ecologists. Although forest inventories often lack the species-level identifications favoured by tropical plant ecologists, their consistency of measurement and vast spatial coverage make them ideally suited for numerical analyses at large scales, and a valuable resource to describe the still poorly understood spatial variation of biomass, diversity, community composition and forest functioning across the South American tropics. Here we show, by using the seven forest inventories complemented with trait and inventory data collected elsewhere, two dominant gradients in tree composition and function across the Amazon, one paralleling a major gradient in soil fertility and the other paralleling a gradient in dry season length. The data set also indicates that the dominance of Fabaceae in the Guiana shield is not necessarily the result of root adaptations to poor soils (nodulation or ectomycorrhizal associations) but perhaps also the result of their remarkably high seed mass there as a potential adaptation to low rates of disturbance.

  9. Formulating a subgrid-scale breakup model for microbubble generation from interfacial collisions

    NASA Astrophysics Data System (ADS)

    Chan, Wai Hong Ronald; Mirjalili, Shahab; Urzay, Javier; Mani, Ali; Moin, Parviz

    2017-11-01

    Multiphase flows often involve impact events that engender important effects like the generation of a myriad of tiny bubbles that are subsequently transported in large liquid bodies. These impact events are created by large-scale phenomena like breaking waves on ocean surfaces, and often involve the relative approach of liquid surfaces. This relative motion generates continuously shrinking length scales as the entrapped gas layer thins and eventually breaks up into microbubbles. The treatment of this disparity in length scales is computationally challenging. In this presentation, a framework is presented that addresses a subgrid-scale (SGS) model aimed at capturing the process of microbubble generation. This work sets up the components in an overarching volume-of-fluid (VoF) toolset and investigates the analytical foundations of an SGS model for describing the breakup of a thin air film trapped between two approaching water bodies in a physical regime corresponding to Mesler entrainment. Constituents of the SGS model, such as the identification of impact events and the accurate computation of the local characteristic curvature in a VoF-based architecture, and the treatment of the air layer breakup, are discussed and illustrated in simplified scenarios. Supported by Office of Naval Research (ONR)/A*STAR (Singapore).

  10. A Scalable Approach for Protein False Discovery Rate Estimation in Large Proteomic Data Sets.

    PubMed

    Savitski, Mikhail M; Wilhelm, Mathias; Hahne, Hannes; Kuster, Bernhard; Bantscheff, Marcus

    2015-09-01

    Calculating the number of confidently identified proteins and estimating false discovery rate (FDR) is a challenge when analyzing very large proteomic data sets such as entire human proteomes. Biological and technical heterogeneity in proteomic experiments further add to the challenge and there are strong differences in opinion regarding the conceptual validity of a protein FDR and no consensus regarding the methodology for protein FDR determination. There are also limitations inherent to the widely used classic target-decoy strategy that particularly show when analyzing very large data sets and that lead to a strong over-representation of decoy identifications. In this study, we investigated the merits of the classic, as well as a novel target-decoy-based protein FDR estimation approach, taking advantage of a heterogeneous data collection comprised of ∼19,000 LC-MS/MS runs deposited in ProteomicsDB (https://www.proteomicsdb.org). The "picked" protein FDR approach treats target and decoy sequences of the same protein as a pair rather than as individual entities and chooses either the target or the decoy sequence depending on which receives the highest score. We investigated the performance of this approach in combination with q-value based peptide scoring to normalize sample-, instrument-, and search engine-specific differences. The "picked" target-decoy strategy performed best when protein scoring was based on the best peptide q-value for each protein yielding a stable number of true positive protein identifications over a wide range of q-value thresholds. We show that this simple and unbiased strategy eliminates a conceptual issue in the commonly used "classic" protein FDR approach that causes overprediction of false-positive protein identification in large data sets. The approach scales from small to very large data sets without losing performance, consistently increases the number of true-positive protein identifications and is readily implemented in proteomics analysis software. © 2015 by The American Society for Biochemistry and Molecular Biology, Inc.

  11. A Scalable Approach for Protein False Discovery Rate Estimation in Large Proteomic Data Sets

    PubMed Central

    Savitski, Mikhail M.; Wilhelm, Mathias; Hahne, Hannes; Kuster, Bernhard; Bantscheff, Marcus

    2015-01-01

    Calculating the number of confidently identified proteins and estimating false discovery rate (FDR) is a challenge when analyzing very large proteomic data sets such as entire human proteomes. Biological and technical heterogeneity in proteomic experiments further add to the challenge and there are strong differences in opinion regarding the conceptual validity of a protein FDR and no consensus regarding the methodology for protein FDR determination. There are also limitations inherent to the widely used classic target–decoy strategy that particularly show when analyzing very large data sets and that lead to a strong over-representation of decoy identifications. In this study, we investigated the merits of the classic, as well as a novel target–decoy-based protein FDR estimation approach, taking advantage of a heterogeneous data collection comprised of ∼19,000 LC-MS/MS runs deposited in ProteomicsDB (https://www.proteomicsdb.org). The “picked” protein FDR approach treats target and decoy sequences of the same protein as a pair rather than as individual entities and chooses either the target or the decoy sequence depending on which receives the highest score. We investigated the performance of this approach in combination with q-value based peptide scoring to normalize sample-, instrument-, and search engine-specific differences. The “picked” target–decoy strategy performed best when protein scoring was based on the best peptide q-value for each protein yielding a stable number of true positive protein identifications over a wide range of q-value thresholds. We show that this simple and unbiased strategy eliminates a conceptual issue in the commonly used “classic” protein FDR approach that causes overprediction of false-positive protein identification in large data sets. The approach scales from small to very large data sets without losing performance, consistently increases the number of true-positive protein identifications and is readily implemented in proteomics analysis software. PMID:25987413

  12. [Adverse Effect Predictions Based on Computational Toxicology Techniques and Large-scale Databases].

    PubMed

    Uesawa, Yoshihiro

    2018-01-01

     Understanding the features of chemical structures related to the adverse effects of drugs is useful for identifying potential adverse effects of new drugs. This can be based on the limited information available from post-marketing surveillance, assessment of the potential toxicities of metabolites and illegal drugs with unclear characteristics, screening of lead compounds at the drug discovery stage, and identification of leads for the discovery of new pharmacological mechanisms. This present paper describes techniques used in computational toxicology to investigate the content of large-scale spontaneous report databases of adverse effects, and it is illustrated with examples. Furthermore, volcano plotting, a new visualization method for clarifying the relationships between drugs and adverse effects via comprehensive analyses, will be introduced. These analyses may produce a great amount of data that can be applied to drug repositioning.

  13. The Middle Miocene of the Fore-Carpathian Basin (Poland, Ukraine and Moldova)

    NASA Astrophysics Data System (ADS)

    Wysocka, Anna; Radwański, Andrzej; Górka, Marcin; Bąbel, Maciej; Radwańska, Urszula; Złotnik, Michał

    2016-09-01

    Studies of Miocene sediments in the Fore-Carpathian Basin, conducted by geologists from the University of Warsaw have provided new insights on the distribution of the facies infilling the basin, particularly in the forebulge and back-bulge zones. The origin of the large-scale sand bodies, evaporitic deposits and large-scale organic buildups is discussed, described and verified. These deposits originated in variable, shallow marine settings, differing in their water chemistry and the dynamics of sedimentary processes, and are unique with regard to the fossil assemblages they yield. Many years of taxonomic, biostratigraphic, palaeoecologic and ecotaphonomic investigations have resulted in the identification of the fossil assemblages of these sediments, their age, sedimentary settings and post-mortem conditions. Detailed studies were focused on corals, polychaetes, most classes of molluscs, crustaceans, echinoderms, and fishes.

  14. Large-eddy simulation of turbulent flow with a surface-mounted two-dimensional obstacle

    NASA Technical Reports Server (NTRS)

    Yang, Kyung-Soo; Ferziger, Joel H.

    1993-01-01

    In this paper, we perform a large eddy simulation (LES) of turbulent flow in a channel containing a two-dimensional obstacle on one wall using a dynamic subgrid-scale model (DSGSM) at Re = 3210, based on bulk velocity above the obstacle and obstacle height; the wall layers are fully resolved. The low Re enables us to perform a DNS (Case 1) against which to validate the LES results. The LES with the DSGSM is designated Case 2. In addition, an LES with the conventional fixed model constant (Case 3) is conducted to allow identification of improvements due to the DSGSM. We also include LES at Re = 82,000 (Case 4) using conventional Smagorinsky subgrid-scale model and a wall-layer model. The results will be compared with the experiment of Dimaczek et al.

  15. Exploration for fossil and nuclear fuels from orbital altitudes

    NASA Technical Reports Server (NTRS)

    Short, N. M.

    1975-01-01

    A review of satellite-based photographic (optical and infrared) and microwave exploration and large-area mapping of the earth's surface in the ERTS program. Synoptic cloud-free coverage of large areas has been achieved with planimetric vertical views of the earth's surface useful in compiling close-to-orthographic mosaics. Radar penetration of cloud cover and infrared penetration of forest cover have been successful to some extent. Geological applications include map editing (with corrections in scale and computer processing of images), landforms analysis, structural geology studies, lithological identification, and exploration for minerals and fuels. Limitations of the method are noted.

  16. Olfactory Performance in a Large Sample of Early-Blind and Late-Blind Individuals.

    PubMed

    Sorokowska, Agnieszka

    2016-10-01

    Previous examinations of olfactory sensitivity in blind people have produced contradictory findings. Thus, whether visual impairment is associated with increased olfactory abilities is unclear. In the present investigation, I aimed to resolve the existing questions via a relatively large-scale study comprising early-blind (N = 43), and late-blind (N = 41) and sighted (N = 84) individuals matched in terms of gender and age. To compare the results with those of previous studies, I combined data from a free odor identification test, extensive psychophysical testing (Sniffin' Sticks test), and self-assessed olfactory performance. The analyses revealed no significant effects of sight on olfactory threshold, odor discrimination, cued identification, or free identification scores; neither was the performance of the early-blind and late-blind participants significantly different. Additionally, the self-assessed olfactory abilities of the blind people were no different than those of the sighted people. These results suggest that sensory compensation in visually impaired is not pronounced with regards to olfactory abilities as measured by standardized smell tests. © The Author 2016. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  17. Application of RNAMlet to surface defect identification of steels

    NASA Astrophysics Data System (ADS)

    Xu, Ke; Xu, Yang; Zhou, Peng; Wang, Lei

    2018-06-01

    As three main production lines of steels, continuous casting slabs, hot rolled steel plates and cold rolled steel strips have different surface appearances and are produced at different speeds of their production lines. Therefore, the algorithms for the surface defect identifications of the three steel products have different requirements for real-time and anti-interference. The existing algorithms cannot be adaptively applied to surface defect identification of the three steel products. A new method of adaptive multi-scale geometric analysis named RNAMlet was proposed. The idea of RNAMlet came from the non-symmetry anti-packing pattern representation model (NAM). The image is decomposed into a set of rectangular blocks asymmetrically according to gray value changes of image pixels. Then two-dimensional Haar wavelet transform is applied to all blocks. If the image background is complex, the number of blocks is large, and more details of the image are utilized. If the image background is simple, the number of blocks is small, and less computation time is needed. RNAMlet was tested with image samples of the three steel products, and compared with three classical methods of multi-scale geometric analysis, including Contourlet, Shearlet and Tetrolet. For the image samples with complicated backgrounds, such as continuous casting slabs and hot rolled steel plates, the defect identification rate obtained by RNAMlet was 1% higher than other three methods. For the image samples with simple backgrounds, such as cold rolled steel strips, the computation time of RNAMlet was one-tenth of the other three MGA methods, while the defect identification rates obtained by RNAMlet were higher than the other three methods.

  18. Large-Scale and Deep Quantitative Proteome Profiling Using Isobaric Labeling Coupled with Two-Dimensional LC-MS/MS

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gritsenko, Marina A.; Xu, Zhe; Liu, Tao

    Comprehensive, quantitative information on abundances of proteins and their post-translational modifications (PTMs) can potentially provide novel biological insights into diseases pathogenesis and therapeutic intervention. Herein, we introduce a quantitative strategy utilizing isobaric stable isotope-labelling techniques combined with two-dimensional liquid chromatography-tandem mass spectrometry (2D-LC-MS/MS) for large-scale, deep quantitative proteome profiling of biological samples or clinical specimens such as tumor tissues. The workflow includes isobaric labeling of tryptic peptides for multiplexed and accurate quantitative analysis, basic reversed-phase LC fractionation and concatenation for reduced sample complexity, and nano-LC coupled to high resolution and high mass accuracy MS analysis for high confidence identification andmore » quantification of proteins. This proteomic analysis strategy has been successfully applied for in-depth quantitative proteomic analysis of tumor samples, and can also be used for integrated proteome and PTM characterization, as well as comprehensive quantitative proteomic analysis across samples from large clinical cohorts.« less

  19. Large-Scale and Deep Quantitative Proteome Profiling Using Isobaric Labeling Coupled with Two-Dimensional LC-MS/MS.

    PubMed

    Gritsenko, Marina A; Xu, Zhe; Liu, Tao; Smith, Richard D

    2016-01-01

    Comprehensive, quantitative information on abundances of proteins and their posttranslational modifications (PTMs) can potentially provide novel biological insights into diseases pathogenesis and therapeutic intervention. Herein, we introduce a quantitative strategy utilizing isobaric stable isotope-labeling techniques combined with two-dimensional liquid chromatography-tandem mass spectrometry (2D-LC-MS/MS) for large-scale, deep quantitative proteome profiling of biological samples or clinical specimens such as tumor tissues. The workflow includes isobaric labeling of tryptic peptides for multiplexed and accurate quantitative analysis, basic reversed-phase LC fractionation and concatenation for reduced sample complexity, and nano-LC coupled to high resolution and high mass accuracy MS analysis for high confidence identification and quantification of proteins. This proteomic analysis strategy has been successfully applied for in-depth quantitative proteomic analysis of tumor samples and can also be used for integrated proteome and PTM characterization, as well as comprehensive quantitative proteomic analysis across samples from large clinical cohorts.

  20. Combined Cycle Engine Large-Scale Inlet for Mode Transition Experiments: System Identification Rack Hardware Design

    NASA Technical Reports Server (NTRS)

    Thomas, Randy; Stueber, Thomas J.

    2013-01-01

    The System Identification (SysID) Rack is a real-time hardware-in-the-loop data acquisition (DAQ) and control instrument rack that was designed and built to support inlet testing in the NASA Glenn Research Center 10- by 10-Foot Supersonic Wind Tunnel. This instrument rack is used to support experiments on the Combined-Cycle Engine Large-Scale Inlet for Mode Transition Experiment (CCE? LIMX). The CCE?LIMX is a testbed for an integrated dual flow-path inlet configuration with the two flow paths in an over-and-under arrangement such that the high-speed flow path is located below the lowspeed flow path. The CCE?LIMX includes multiple actuators that are designed to redirect airflow from one flow path to the other; this action is referred to as "inlet mode transition." Multiple phases of experiments have been planned to support research that investigates inlet mode transition: inlet characterization (Phase-1) and system identification (Phase-2). The SysID Rack hardware design met the following requirements to support Phase-1 and Phase-2 experiments: safely and effectively move multiple actuators individually or synchronously; sample and save effector control and position sensor feedback signals; automate control of actuator positioning based on a mode transition schedule; sample and save pressure sensor signals; and perform DAQ and control processes operating at 2.5 KHz. This document describes the hardware components used to build the SysID Rack including their function, specifications, and system interface. Furthermore, provided in this document are a SysID Rack effectors signal list (signal flow); system identification experiment setup; illustrations indicating a typical SysID Rack experiment; and a SysID Rack performance overview for Phase-1 and Phase-2 experiments. The SysID Rack described in this document was a useful tool to meet the project objectives.

  1. Protein-Protein Interactions in a Crowded Environment: An Analysis via Cross-Docking Simulations and Evolutionary Information

    PubMed Central

    Lopes, Anne; Sacquin-Mora, Sophie; Dimitrova, Viktoriya; Laine, Elodie; Ponty, Yann; Carbone, Alessandra

    2013-01-01

    Large-scale analyses of protein-protein interactions based on coarse-grain molecular docking simulations and binding site predictions resulting from evolutionary sequence analysis, are possible and realizable on hundreds of proteins with variate structures and interfaces. We demonstrated this on the 168 proteins of the Mintseris Benchmark 2.0. On the one hand, we evaluated the quality of the interaction signal and the contribution of docking information compared to evolutionary information showing that the combination of the two improves partner identification. On the other hand, since protein interactions usually occur in crowded environments with several competing partners, we realized a thorough analysis of the interactions of proteins with true partners but also with non-partners to evaluate whether proteins in the environment, competing with the true partner, affect its identification. We found three populations of proteins: strongly competing, never competing, and interacting with different levels of strength. Populations and levels of strength are numerically characterized and provide a signature for the behavior of a protein in the crowded environment. We showed that partner identification, to some extent, does not depend on the competing partners present in the environment, that certain biochemical classes of proteins are intrinsically easier to analyze than others, and that small proteins are not more promiscuous than large ones. Our approach brings to light that the knowledge of the binding site can be used to reduce the high computational cost of docking simulations with no consequence in the quality of the results, demonstrating the possibility to apply coarse-grain docking to datasets made of thousands of proteins. Comparison with all available large-scale analyses aimed to partner predictions is realized. We release the complete decoys set issued by coarse-grain docking simulations of both true and false interacting partners, and their evolutionary sequence analysis leading to binding site predictions. Download site: http://www.lgm.upmc.fr/CCDMintseris/ PMID:24339765

  2. Adaptive Identification and Characterization of Polar Ionization Patches

    NASA Technical Reports Server (NTRS)

    Coley, W. R.; Heelis, R. A.

    1995-01-01

    Dynamics Explorer 2 (DE 2) spacecraft data are used to detect and characterize polar cap 'ionization patches' loosely defined as large-scale (greater than 100 km) regions where the F region plasma density is significantly enhanced (approx greater than 100%) above the background level. These patches are generally believed to develop in or equatorward of the dayside cusp region and then drift in an antisunward direction over the polar cap. We have developed a flexible algorithm for the identification and characterization of these structures, as a function of scale-size and density enhancement, using data from the retarding potential analyzer, the ion drift meter, and the langmuir probe on board the DE 2 satellite. This algorithm was used to study the structure and evolution of ionization patches as they cross the polar cap. The results indicate that in the altitude region from 240 to 950 km ion density enhancements greater than a factor of 3 above the background level are relatively rare. Further, the ionization patches show a preferred horizontal scale size of 300-400 km. There exists a clear seasonal and universal time dependence to the occurrence frequency of patches with a northern hemisphere maximum centered on the winter solstice and the 1200-2000 UT interval.

  3. MRMPROBS: a data assessment and metabolite identification tool for large-scale multiple reaction monitoring based widely targeted metabolomics.

    PubMed

    Tsugawa, Hiroshi; Arita, Masanori; Kanazawa, Mitsuhiro; Ogiwara, Atsushi; Bamba, Takeshi; Fukusaki, Eiichiro

    2013-05-21

    We developed a new software program, MRMPROBS, for widely targeted metabolomics by using the large-scale multiple reaction monitoring (MRM) mode. The strategy became increasingly popular for the simultaneous analysis of up to several hundred metabolites at high sensitivity, selectivity, and quantitative capability. However, the traditional method of assessing measured metabolomics data without probabilistic criteria is not only time-consuming but is often subjective and makeshift work. Our program overcomes these problems by detecting and identifying metabolites automatically, by separating isomeric metabolites, and by removing background noise using a probabilistic score defined as the odds ratio from an optimized multivariate logistic regression model. Our software program also provides a user-friendly graphical interface to curate and organize data matrices and to apply principal component analyses and statistical tests. For a demonstration, we conducted a widely targeted metabolome analysis (152 metabolites) of propagating Saccharomyces cerevisiae measured at 15 time points by gas and liquid chromatography coupled to triple quadrupole mass spectrometry. MRMPROBS is a useful and practical tool for the assessment of large-scale MRM data available to any instrument or any experimental condition.

  4. 49 CFR 178.905 - Large Packaging identification codes.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 49 Transportation 2 2010-10-01 2010-10-01 false Large Packaging identification codes. 178.905... FOR PACKAGINGS Large Packagings Standards § 178.905 Large Packaging identification codes. Large packaging code designations consist of: two numerals specified in paragraph (a) of this section; followed by...

  5. Retinal identification based on an Improved Circular Gabor Filter and Scale Invariant Feature Transform.

    PubMed

    Meng, Xianjing; Yin, Yilong; Yang, Gongping; Xi, Xiaoming

    2013-07-18

    Retinal identification based on retinal vasculatures in the retina provides the most secure and accurate means of authentication among biometrics and has primarily been used in combination with access control systems at high security facilities. Recently, there has been much interest in retina identification. As digital retina images always suffer from deformations, the Scale Invariant Feature Transform (SIFT), which is known for its distinctiveness and invariance for scale and rotation, has been introduced to retinal based identification. However, some shortcomings like the difficulty of feature extraction and mismatching exist in SIFT-based identification. To solve these problems, a novel preprocessing method based on the Improved Circular Gabor Transform (ICGF) is proposed. After further processing by the iterated spatial anisotropic smooth method, the number of uninformative SIFT keypoints is decreased dramatically. Tested on the VARIA and eight simulated retina databases combining rotation and scaling, the developed method presents promising results and shows robustness to rotations and scale changes.

  6. Retinal Identification Based on an Improved Circular Gabor Filter and Scale Invariant Feature Transform

    PubMed Central

    Meng, Xianjing; Yin, Yilong; Yang, Gongping; Xi, Xiaoming

    2013-01-01

    Retinal identification based on retinal vasculatures in the retina provides the most secure and accurate means of authentication among biometrics and has primarily been used in combination with access control systems at high security facilities. Recently, there has been much interest in retina identification. As digital retina images always suffer from deformations, the Scale Invariant Feature Transform (SIFT), which is known for its distinctiveness and invariance for scale and rotation, has been introduced to retinal based identification. However, some shortcomings like the difficulty of feature extraction and mismatching exist in SIFT-based identification. To solve these problems, a novel preprocessing method based on the Improved Circular Gabor Transform (ICGF) is proposed. After further processing by the iterated spatial anisotropic smooth method, the number of uninformative SIFT keypoints is decreased dramatically. Tested on the VARIA and eight simulated retina databases combining rotation and scaling, the developed method presents promising results and shows robustness to rotations and scale changes. PMID:23873409

  7. Assessing Bodily Preoccupations is sufficient: clinically effective screening for hypochondriasis.

    PubMed

    Höfling, Volkmar; Weck, Florian

    2013-12-01

    Hypochondriasis is a persistent psychiatric disorder and is associated with increased utilisation of health care services. However, effective psychiatric consultation interventions and CBT treatments are available. In the present study, we provide evidence of clinically effective screening for hypochondriasis. We describe the clinically effective identification of patients with a high probability of suffering from hypochondriasis. This identification is achieved by means of two brief standardised screening instruments, namely the Bodily Preoccupation (BP) Scale with 3 items and the Whiteley-7 (WI-7) with 7 items. Both the BP scale and the WI-7 were examined in a sample of 228 participants (72 with hypochondriasis, 80 with anxiety disorders and 76 healthy controls) in a large psychotherapy outpatients' unit, applying the DSM-IV criteria. Cut-off values for the BP scale and the WI-7 were computed to identify patients with a high probability of suffering from hypochondriasis. Additionally, other self-report symptom severity scales were completed in order to examine discriminant and convergent validity. Data was collected from June 2010 to March 2013. The BP scale and the WI-7 discriminated significantly between patients with hypochondriasis and those with an anxiety disorder (d=2.42 and d=2.34). Cut-off values for these two screening scales could be provided, thus identifying patients with a high probability of suffering from hypochondriasis. In order to reduce costs, the BP scale or the WI-7 should be applied in medical or primary care settings, to screen for patients with a high probability of hypochondriasis and to transfer them to further assessment and effective treatment. © 2013.

  8. Identification of possible non-stationary effects in a new type of vortex furnace

    NASA Astrophysics Data System (ADS)

    Shadrin, Evgeniy Yu.; Anufriev, Igor S.; Papulov, Anatoly P.

    2017-10-01

    The article presents the results of an experimental study of pressure and velocity pulsations in the model of improved vortex furnace with distributed air supply and vertically oriented nozzles of the secondary blast. Investigation of aerodynamic characteristics of a swirling flow with different regime parameters was conducted in an isothermal laboratory model (in 1:25 scale) of vortex furnace using laser Doppler measuring system and pressure pulsations analyzer. The obtained results have revealed a number of features of the flow structure, and the spectral analysis of pressure and velocity pulsations allows to speak about the absence of large-scale unsteady vortical structures in the studied design.

  9. The influence of false color infrared display on training field identification. [for crop inventories

    NASA Technical Reports Server (NTRS)

    Coberly, W. A.; Tubbs, J. D.; Odell, P. L.

    1979-01-01

    The overall success of large-scale crop inventories of agricultural regions using Landsat multispectral scanner data is highly dependent upon the labeling of training data by analyst/photointerpreters. The principal analyst tool in labeling training data is a false color infrared composite of Landsat bands 4, 5, and 7. In this paper, this color display is investigated and its influence upon classification errors is partially determined.

  10. Identification of Variant-Specific Functions of PIK3CA by Rapid Phenotyping of Rare Mutations | Office of Cancer Genomics

    Cancer.gov

    Large-scale sequencing efforts are uncovering the complexity of cancer genomes, which are composed of causal "driver" mutations that promote tumor progression along with many more pathologically neutral "passenger" events. The majority of mutations, both in known cancer drivers and uncharacterized genes, are generally of low occurrence, highlighting the need to functionally annotate the long tail of infrequent mutations present in heterogeneous cancers.

  11. Large Scale Single Nucleotide Polymorphism Study of PD Susceptibility

    DTIC Science & Technology

    2006-03-01

    familial PD, the results of intensive investigations of polymorphisms in dozens of genes related to sporadic, late onset, typical PD have not shown...association between classical, sporadic PD and 2386 SNPs in 23 genes implicated in the pathogenesis of PD; (2) construct haplotypes based on the SNP...derived from this study may be applied in other complex disorders for the identification of susceptibility genes , as well as in genome-wide SNP

  12. In the eye of the beholder: the effect of rater variability and different rating scales on QTL mapping.

    PubMed

    Poland, Jesse A; Nelson, Rebecca J

    2011-02-01

    The agronomic importance of developing durably resistant cultivars has led to substantial research in the field of quantitative disease resistance (QDR) and, in particular, mapping quantitative trait loci (QTL) for disease resistance. The assessment of QDR is typically conducted by visual estimation of disease severity, which raises concern over the accuracy and precision of visual estimates. Although previous studies have examined the factors affecting the accuracy and precision of visual disease assessment in relation to the true value of disease severity, the impact of this variability on the identification of disease resistance QTL has not been assessed. In this study, the effects of rater variability and rating scales on mapping QTL for northern leaf blight resistance in maize were evaluated in a recombinant inbred line population grown under field conditions. The population of 191 lines was evaluated by 22 different raters using a direct percentage estimate, a 0-to-9 ordinal rating scale, or both. It was found that more experienced raters had higher precision and that using a direct percentage estimation of diseased leaf area produced higher precision than using an ordinal scale. QTL mapping was then conducted using the disease estimates from each rater using stepwise general linear model selection (GLM) and inclusive composite interval mapping (ICIM). For GLM, the same QTL were largely found across raters, though some QTL were only identified by a subset of raters. The magnitudes of estimated allele effects at identified QTL varied drastically, sometimes by as much as threefold. ICIM produced highly consistent results across raters and for the different rating scales in identifying the location of QTL. We conclude that, despite variability between raters, the identification of QTL was largely consistent among raters, particularly when using ICIM. However, care should be taken in estimating QTL allele effects, because this was highly variable and rater dependent.

  13. Finding the needle in a haystack: identification of cases of Lynch syndrome with MLH1 epimutation.

    PubMed

    Hitchins, Megan P

    2016-07-01

    Constitutional epimutation of the DNA mismatch repair gene, MLH1, represents a minor cause of Lynch syndrome. MLH1 epimutations are characterized by the soma-wide distribution of methylation of a single allele of the MLH1 promoter accompanied by constitutive allelic loss of transcription. 'Primary' MLH1 epimutations, considered pure epigenetic defects, tend to arise de novo in patients without a family history or any apparent genetic mutation. These demonstrate non-Mendelian inheritance. 'Secondary' MLH1 epimutations have a genetic basis and have been linked to non-coding genetic alterations in the vicinity of MLH1. These demonstrate autosomal dominant inheritance. Despite convincing evidence of their role in causing Lynch-type cancers, routine screening for MLH1 epimutations has not been widely adopted. Complicating factors may include: the need to perform additional methylation-based testing beyond the standard genetic screening for a germline mutation; the lack of a consensus algorithm for the selection of patients warranting MLH1 epimutation testing; overlapping molecular pathology features of MLH1 methylation and loss of MLH1 expression with more prevalent sporadic MSI cancers; the rarity of MLH1 epimutation; the variable inter-generational inheritance patterns; and the cost-effectiveness of screening. Nevertheless, a positive molecular diagnosis of MLH1 epimutation is clinically important because carriers have a high personal risk of developing metachronous Lynch-type cancers, and their relatives may also be at risk of carriage. Extending existing universal and clinic-based screening algorithms for Lynch syndrome to include an additional arm of selection criteria for cases warranting MLH1 epimutation testing could provide a cost-effective means of diagnosing these cases.

  14. DNA barcodes reveal microevolutionary signals in fire response trait in two legume genera

    PubMed Central

    Bello, Abubakar; Daru, Barnabas H.; Stirton, Charles H.; Chimphango, Samson B. M.; van der Bank, Michelle; Maurin, Olivier; Muasya, A. Muthama

    2015-01-01

    Large-scale DNA barcoding provides a new technique for species identification and evaluation of relationships across various levels (populations and species) and may reveal fundamental processes in recently diverged species. Here, we analysed DNA sequence variation in the recently diverged legumes from the Psoraleeae (Fabaceae) occurring in the Cape Floristic Region (CFR) of southern Africa to test the utility of DNA barcodes in species identification and discrimination. We further explored the phylogenetic signal on fire response trait (reseeding and resprouting) at species and generic levels. We showed that Psoraleoid legumes of the CFR exhibit a barcoding gap yielding the combination of matK and rbcLa (matK + rbcLa) data set as a better barcode than single regions. We found a high score (100 %) of correct identification of individuals to their respective genera but a very low score (<50 %) in identifying them to species. We found a considerable match (54 %) between genetic species and morphologically delimited species. We also found that different lineages showed a weak but significant phylogenetic conservatism in their response to fire as reseeders or resprouters, with more clustering of resprouters than would be expected by chance. These novel microevolutionary patterns might be acting continuously over time to produce multi-scale regularities of biodiversity. This study provides the first insight into the DNA barcoding campaign of land plants in species identification and detection of the phylogenetic signal in recently diverged lineages of the CFR. PMID:26507570

  15. Characterizing the replicability of cell types defined by single cell RNA-sequencing data using MetaNeighbor.

    PubMed

    Crow, Megan; Paul, Anirban; Ballouz, Sara; Huang, Z Josh; Gillis, Jesse

    2018-02-28

    Single-cell RNA-sequencing (scRNA-seq) technology provides a new avenue to discover and characterize cell types; however, the experiment-specific technical biases and analytic variability inherent to current pipelines may undermine its replicability. Meta-analysis is further hampered by the use of ad hoc naming conventions. Here we demonstrate our replication framework, MetaNeighbor, that quantifies the degree to which cell types replicate across datasets, and enables rapid identification of clusters with high similarity. We first measure the replicability of neuronal identity, comparing results across eight technically and biologically diverse datasets to define best practices for more complex assessments. We then apply this to novel interneuron subtypes, finding that 24/45 subtypes have evidence of replication, which enables the identification of robust candidate marker genes. Across tasks we find that large sets of variably expressed genes can identify replicable cell types with high accuracy, suggesting a general route forward for large-scale evaluation of scRNA-seq data.

  16. Large-scale identification of chemically induced mutations in Drosophila melanogaster

    PubMed Central

    Haelterman, Nele A.; Jiang, Lichun; Li, Yumei; Bayat, Vafa; Sandoval, Hector; Ugur, Berrak; Tan, Kai Li; Zhang, Ke; Bei, Danqing; Xiong, Bo; Charng, Wu-Lin; Busby, Theodore; Jawaid, Adeel; David, Gabriela; Jaiswal, Manish; Venken, Koen J.T.; Yamamoto, Shinya

    2014-01-01

    Forward genetic screens using chemical mutagens have been successful in defining the function of thousands of genes in eukaryotic model organisms. The main drawback of this strategy is the time-consuming identification of the molecular lesions causative of the phenotypes of interest. With whole-genome sequencing (WGS), it is now possible to sequence hundreds of strains, but determining which mutations are causative among thousands of polymorphisms remains challenging. We have sequenced 394 mutant strains, generated in a chemical mutagenesis screen, for essential genes on the Drosophila X chromosome and describe strategies to reduce the number of candidate mutations from an average of ∼3500 to 35 single-nucleotide variants per chromosome. By combining WGS with a rough mapping method based on large duplications, we were able to map 274 (∼70%) mutations. We show that these mutations are causative, using small 80-kb duplications that rescue lethality. Hence, our findings demonstrate that combining rough mapping with WGS dramatically expands the toolkit necessary for assigning function to genes. PMID:25258387

  17. Cross-Identification of Astronomical Catalogs on Multiple GPUs

    NASA Astrophysics Data System (ADS)

    Lee, M. A.; Budavári, T.

    2013-10-01

    One of the most fundamental problems in observational astronomy is the cross-identification of sources. Observations are made in different wavelengths, at different times, and from different locations and instruments, resulting in a large set of independent observations. The scientific outcome is often limited by our ability to quickly perform meaningful associations between detections. The matching, however, is difficult scientifically, statistically, as well as computationally. The former two require detailed physical modeling and advanced probabilistic concepts; the latter is due to the large volumes of data and the problem's combinatorial nature. In order to tackle the computational challenge and to prepare for future surveys, whose measurements will be exponentially increasing in size past the scale of feasible CPU-based solutions, we developed a new implementation which addresses the issue by performing the associations on multiple Graphics Processing Units (GPUs). Our implementation utilizes up to 6 GPUs in combination with the Thrust library to achieve an over 40x speed up verses the previous best implementation running on a multi-CPU SQL Server.

  18. Large-scale time-lapse microscopy of Oct4 expression in human embryonic stem cell colonies.

    PubMed

    Bhadriraju, Kiran; Halter, Michael; Amelot, Julien; Bajcsy, Peter; Chalfoun, Joe; Vandecreme, Antoine; Mallon, Barbara S; Park, Kye-Yoon; Sista, Subhash; Elliott, John T; Plant, Anne L

    2016-07-01

    Identification and quantification of the characteristics of stem cell preparations is critical for understanding stem cell biology and for the development and manufacturing of stem cell based therapies. We have developed image analysis and visualization software that allows effective use of time-lapse microscopy to provide spatial and dynamic information from large numbers of human embryonic stem cell colonies. To achieve statistically relevant sampling, we examined >680 colonies from 3 different preparations of cells over 5days each, generating a total experimental dataset of 0.9 terabyte (TB). The 0.5 Giga-pixel images at each time point were represented by multi-resolution pyramids and visualized using the Deep Zoom Javascript library extended to support viewing Giga-pixel images over time and extracting data on individual colonies. We present a methodology that enables quantification of variations in nominally-identical preparations and between colonies, correlation of colony characteristics with Oct4 expression, and identification of rare events. Copyright © 2016. Published by Elsevier B.V.

  19. A large-scale cryoelectronic system for biological sample banking

    NASA Astrophysics Data System (ADS)

    Shirley, Stephen G.; Durst, Christopher H. P.; Fuchs, Christian C.; Zimmermann, Heiko; Ihmig, Frank R.

    2009-11-01

    We describe a polymorphic electronic infrastructure for managing biological samples stored over liquid nitrogen. As part of this system we have developed new cryocontainers and carrier plates attached to Flash memory chips to have a redundant and portable set of data at each sample. Our experimental investigations show that basic Flash operation and endurance is adequate for the application down to liquid nitrogen temperatures. This identification technology can provide the best sample identification, documentation and tracking that brings added value to each sample. The first application of the system is in a worldwide collaborative research towards the production of an AIDS vaccine. The functionality and versatility of the system can lead to an essential optimization of sample and data exchange for global clinical studies.

  20. Identification and Analysis of Antiviral Compounds Against Poliovirus.

    PubMed

    Leyssen, Pieter; Franco, David; Tijsma, Aloys; Lacroix, Céline; De Palma, Armando; Neyts, Johan

    2016-01-01

    The Global Polio Eradication Initiative, launched in 1988, had as its goal the eradication of polio worldwide by the year 2000 through large-scale vaccinations campaigns with the live attenuated oral PV vaccine (OPV) (Griffiths et al., Biologicals 34:73-74, 2006). Despite substantial progress, polio remains endemic in several countries and new imported cases are reported on a regular basis ( http://www.polioeradication.org/casecount.asp ).It was recognized by the poliovirus research community that developing antivirals against poliovirus would be invaluable in the post-OPV era. Here, we describe three methods essential for the identification of selective inhibitors of poliovirus replication and for determining their mode of action by time-of-drug-addition studies as well as by the isolation of compound-resistant poliovirus variants.

  1. Design of a large-scale femtoliter droplet array for single-cell analysis of drug-tolerant and drug-resistant bacteria.

    PubMed

    Iino, Ryota; Matsumoto, Yoshimi; Nishino, Kunihiko; Yamaguchi, Akihito; Noji, Hiroyuki

    2013-01-01

    Single-cell analysis is a powerful method to assess the heterogeneity among individual cells, enabling the identification of very rare cells with properties that differ from those of the majority. In this Methods Article, we describe the use of a large-scale femtoliter droplet array to enclose, isolate, and analyze individual bacterial cells. As a first example, we describe the single-cell detection of drug-tolerant persisters of Pseudomonas aeruginosa treated with the antibiotic carbenicillin. As a second example, this method was applied to the single-cell evaluation of drug efflux activity, which causes acquired antibiotic resistance of bacteria. The activity of the MexAB-OprM multidrug efflux pump system from Pseudomonas aeruginosa was expressed in Escherichia coli and the effect of an inhibitor D13-9001 were assessed at the single cell level.

  2. Mother Nature versus human nature: public compliance with evacuation and quarantine.

    PubMed

    Manuell, Mary-Elise; Cukor, Jeffrey

    2011-04-01

    Effectively controlling the spread of contagious illnesses has become a critical focus of disaster planning. It is likely that quarantine will be a key part of the overall public health strategy utilised during a pandemic, an act of bioterrorism or other emergencies involving contagious agents. While the United States lacks recent experience of large-scale quarantines, it has considerable accumulated experience of large-scale evacuations. Risk perception, life circumstance, work-related issues, and the opinions of influential family, friends and credible public spokespersons all play a role in determining compliance with an evacuation order. Although the comparison is not reported elsewhere to our knowledge, this review of the principal factors affecting compliance with evacuations demonstrates many similarities with those likely to occur during a quarantine. Accurate identification and understanding of barriers to compliance allows for improved planning to protect the public more effectively. © 2011 The Author(s). Disasters © Overseas Development Institute, 2011.

  3. Ubiquitinated Proteome: Ready for Global?*

    PubMed Central

    Shi, Yi; Xu, Ping; Qin, Jun

    2011-01-01

    Ubiquitin (Ub) is a small and highly conserved protein that can covalently modify protein substrates. Ubiquitination is one of the major post-translational modifications that regulate a broad spectrum of cellular functions. The advancement of mass spectrometers as well as the development of new affinity purification tools has greatly expedited proteome-wide analysis of several post-translational modifications (e.g. phosphorylation, glycosylation, and acetylation). In contrast, large-scale profiling of lysine ubiquitination remains a challenge. Most recently, new Ub affinity reagents such as Ub remnant antibody and tandem Ub binding domains have been developed, allowing for relatively large-scale detection of several hundreds of lysine ubiquitination events in human cells. Here we review different strategies for the identification of ubiquitination site and discuss several issues associated with data analysis. We suggest that careful interpretation and orthogonal confirmation of MS spectra is necessary to minimize false positive assignments by automatic searching algorithms. PMID:21339389

  4. MRMPROBS suite for metabolomics using large-scale MRM assays.

    PubMed

    Tsugawa, Hiroshi; Kanazawa, Mitsuhiro; Ogiwara, Atsushi; Arita, Masanori

    2014-08-15

    We developed new software environment for the metabolome analysis of large-scale multiple reaction monitoring (MRM) assays. It supports the data format of four major mass spectrometer vendors and mzML common data format. This program provides a process pipeline from the raw-format import to high-dimensional statistical analyses. The novel aspect is graphical user interface-based visualization to perform peak quantification, to interpolate missing values and to normalize peaks interactively based on quality control samples. Together with the software platform, the MRM standard library of 301 metabolites with 775 transitions is also available, which contributes to the reliable peak identification by using retention time and ion abundances. MRMPROBS is available for Windows OS under the creative-commons by-attribution license at http://prime.psc.riken.jp. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  5. Promoting R & D in photobiological hydrogen production utilizing mariculture-raised cyanobacteria.

    PubMed

    Sakurai, Hidehiro; Masukawa, Hajime

    2007-01-01

    This review article explores the potential of using mariculture-raised cyanobacteria as solar energy converters of hydrogen (H(2)). The exploitation of the sea surface for large-scale renewable energy production and the reasons for selecting the economical, nitrogenase-based systems of cyanobacteria for H(2) production, are described in terms of societal benefits. Reports of cyanobacterial photobiological H(2) production are summarized with respect to specific activity, efficiency of solar energy conversion, and maximum H(2) concentration attainable. The need for further improvements in biological parameters such as low-light saturation properties, sustainability of H(2) production, and so forth, and the means to overcome these difficulties through the identification of promising wild-type strains followed by optimization of the selected strains using genetic engineering are also discussed. Finally, a possible mechanism for the development of economical large-scale mariculture operations in conjunction with international cooperation and social acceptance is outlined.

  6. Cloud-based solution to identify statistically significant MS peaks differentiating sample categories.

    PubMed

    Ji, Jun; Ling, Jeffrey; Jiang, Helen; Wen, Qiaojun; Whitin, John C; Tian, Lu; Cohen, Harvey J; Ling, Xuefeng B

    2013-03-23

    Mass spectrometry (MS) has evolved to become the primary high throughput tool for proteomics based biomarker discovery. Until now, multiple challenges in protein MS data analysis remain: large-scale and complex data set management; MS peak identification, indexing; and high dimensional peak differential analysis with the concurrent statistical tests based false discovery rate (FDR). "Turnkey" solutions are needed for biomarker investigations to rapidly process MS data sets to identify statistically significant peaks for subsequent validation. Here we present an efficient and effective solution, which provides experimental biologists easy access to "cloud" computing capabilities to analyze MS data. The web portal can be accessed at http://transmed.stanford.edu/ssa/. Presented web application supplies large scale MS data online uploading and analysis with a simple user interface. This bioinformatic tool will facilitate the discovery of the potential protein biomarkers using MS.

  7. Enrichment and separation techniques for large-scale proteomics analysis of the protein post-translational modifications.

    PubMed

    Huang, Junfeng; Wang, Fangjun; Ye, Mingliang; Zou, Hanfa

    2014-11-06

    Comprehensive analysis of the post-translational modifications (PTMs) on proteins at proteome level is crucial to elucidate the regulatory mechanisms of various biological processes. In the past decades, thanks to the development of specific PTM enrichment techniques and efficient multidimensional liquid chromatography (LC) separation strategy, the identification of protein PTMs have made tremendous progress. A huge number of modification sites for some major protein PTMs have been identified by proteomics analysis. In this review, we first introduced the recent progresses of PTM enrichment methods for the analysis of several major PTMs including phosphorylation, glycosylation, ubiquitination, acetylation, methylation, and oxidation/reduction status. We then briefly summarized the challenges for PTM enrichment. Finally, we introduced the fractionation and separation techniques for efficient separation of PTM peptides in large-scale PTM analysis. Copyright © 2014 Elsevier B.V. All rights reserved.

  8. Listening to the Deep: live monitoring of ocean noise and cetacean acoustic signals.

    PubMed

    André, M; van der Schaar, M; Zaugg, S; Houégnigan, L; Sánchez, A M; Castell, J V

    2011-01-01

    The development and broad use of passive acoustic monitoring techniques have the potential to help assessing the large-scale influence of artificial noise on marine organisms and ecosystems. Deep-sea observatories have the potential to play a key role in understanding these recent acoustic changes. LIDO (Listening to the Deep Ocean Environment) is an international project that is allowing the real-time long-term monitoring of marine ambient noise as well as marine mammal sounds at cabled and standalone observatories. Here, we present the overall development of the project and the use of passive acoustic monitoring (PAM) techniques to provide the scientific community with real-time data at large spatial and temporal scales. Special attention is given to the extraction and identification of high frequency cetacean echolocation signals given the relevance of detecting target species, e.g. beaked whales, in mitigation processes, e.g. during military exercises. Copyright © 2011. Published by Elsevier Ltd.

  9. A Normalization-Free and Nonparametric Method Sharpens Large-Scale Transcriptome Analysis and Reveals Common Gene Alteration Patterns in Cancers.

    PubMed

    Li, Qi-Gang; He, Yong-Han; Wu, Huan; Yang, Cui-Ping; Pu, Shao-Yan; Fan, Song-Qing; Jiang, Li-Ping; Shen, Qiu-Shuo; Wang, Xiao-Xiong; Chen, Xiao-Qiong; Yu, Qin; Li, Ying; Sun, Chang; Wang, Xiangting; Zhou, Jumin; Li, Hai-Peng; Chen, Yong-Bin; Kong, Qing-Peng

    2017-01-01

    Heterogeneity in transcriptional data hampers the identification of differentially expressed genes (DEGs) and understanding of cancer, essentially because current methods rely on cross-sample normalization and/or distribution assumption-both sensitive to heterogeneous values. Here, we developed a new method, Cross-Value Association Analysis (CVAA), which overcomes the limitation and is more robust to heterogeneous data than the other methods. Applying CVAA to a more complex pan-cancer dataset containing 5,540 transcriptomes discovered numerous new DEGs and many previously rarely explored pathways/processes; some of them were validated, both in vitro and in vivo , to be crucial in tumorigenesis, e.g., alcohol metabolism ( ADH1B ), chromosome remodeling ( NCAPH ) and complement system ( Adipsin ). Together, we present a sharper tool to navigate large-scale expression data and gain new mechanistic insights into tumorigenesis.

  10. Cell line name recognition in support of the identification of synthetic lethality in cancer from text

    PubMed Central

    Kaewphan, Suwisa; Van Landeghem, Sofie; Ohta, Tomoko; Van de Peer, Yves; Ginter, Filip; Pyysalo, Sampo

    2016-01-01

    Motivation: The recognition and normalization of cell line names in text is an important task in biomedical text mining research, facilitating for instance the identification of synthetically lethal genes from the literature. While several tools have previously been developed to address cell line recognition, it is unclear whether available systems can perform sufficiently well in realistic and broad-coverage applications such as extracting synthetically lethal genes from the cancer literature. In this study, we revisit the cell line name recognition task, evaluating both available systems and newly introduced methods on various resources to obtain a reliable tagger not tied to any specific subdomain. In support of this task, we introduce two text collections manually annotated for cell line names: the broad-coverage corpus Gellus and CLL, a focused target domain corpus. Results: We find that the best performance is achieved using NERsuite, a machine learning system based on Conditional Random Fields, trained on the Gellus corpus and supported with a dictionary of cell line names. The system achieves an F-score of 88.46% on the test set of Gellus and 85.98% on the independently annotated CLL corpus. It was further applied at large scale to 24 302 102 unannotated articles, resulting in the identification of 5 181 342 cell line mentions, normalized to 11 755 unique cell line database identifiers. Availability and implementation: The manually annotated datasets, the cell line dictionary, derived corpora, NERsuite models and the results of the large-scale run on unannotated texts are available under open licenses at http://turkunlp.github.io/Cell-line-recognition/. Contact: sukaew@utu.fi PMID:26428294

  11. Identification and estimation of the area planted with irrigated rice based on the visual interpretation of LANDSAT MSS data

    NASA Technical Reports Server (NTRS)

    Parada, N. D. J. (Principal Investigator); Moreira, M. A.; Assuncao, G. V.; Novaes, R. A.; Mendoza, A. A. B.; Bauer, C. A.; Ritter, I. T.; Barros, J. A. I.; Perez, J. E.; Thedy, J. L. O.

    1983-01-01

    The objective was to test the feasibility of the application of MSS-LANDSAT data to irrigated rice crop identification and area evaluation, within four rice growing regions of the Rio Grande do Sul state, in order to extend the methodology for the whole state. The applied methodology was visual interpretation of the following LANDSAT products: channels 5 and 7 black and white imageries and color infrared composite imageries all at the scale of 1:250.000. For crop identification and evaluation, the multispectral criterion and the seasonal variation were utilized. Based on the results it was possible to conclude that: (1) the satellite data were efficient for crop area identification and evaluation; (2) the utilization of the multispectral criterion, allied to the seasonal variation of the rice crop areas from the other crops and, (3) the large cloud cover percentage found in the satellite data made it impossible to realize a rice crop spectral monitoring and, therefore, to define the best dates for such data acquisition for rice crop assessment.

  12. Looking back on a decade of barcoding crustaceans

    PubMed Central

    Raupach, Michael J.; Radulovici, Adriana E.

    2015-01-01

    Abstract Species identification represents a pivotal component for large-scale biodiversity studies and conservation planning but represents a challenge for many taxa when using morphological traits only. Consequently, alternative identification methods based on molecular markers have been proposed. In this context, DNA barcoding has become a popular and accepted method for the identification of unknown animals across all life stages by comparison to a reference library. In this review we examine the progress of barcoding studies for the Crustacea using the Web of Science data base from 2003 to 2014. All references were classified in terms of taxonomy covered, subject area (identification/library, genetic variability, species descriptions, phylogenetics, methods, pseudogenes/numts), habitat, geographical area, authors, journals, citations, and the use of the Barcode of Life Data Systems (BOLD). Our analysis revealed a total number of 164 barcoding studies for crustaceans with a preference for malacostracan crustaceans, in particular Decapoda, and for building reference libraries in order to identify organisms. So far, BOLD did not establish itself as a popular informatics platform among carcinologists although it offers many advantages for standardized data storage, analyses and publication. PMID:26798245

  13. Energy scaling and reduction in controlling complex networks

    PubMed Central

    Chen, Yu-Zhong; Wang, Le-Zhi; Wang, Wen-Xu; Lai, Ying-Cheng

    2016-01-01

    Recent works revealed that the energy required to control a complex network depends on the number of driving signals and the energy distribution follows an algebraic scaling law. If one implements control using a small number of drivers, e.g. as determined by the structural controllability theory, there is a high probability that the energy will diverge. We develop a physical theory to explain the scaling behaviour through identification of the fundamental structural elements, the longest control chains (LCCs), that dominate the control energy. Based on the LCCs, we articulate a strategy to drastically reduce the control energy (e.g. in a large number of real-world networks). Owing to their structural nature, the LCCs may shed light on energy issues associated with control of nonlinear dynamical networks. PMID:27152220

  14. Large-scale DCMs for resting-state fMRI.

    PubMed

    Razi, Adeel; Seghier, Mohamed L; Zhou, Yuan; McColgan, Peter; Zeidman, Peter; Park, Hae-Jeong; Sporns, Olaf; Rees, Geraint; Friston, Karl J

    2017-01-01

    This paper considers the identification of large directed graphs for resting-state brain networks based on biophysical models of distributed neuronal activity, that is, effective connectivity . This identification can be contrasted with functional connectivity methods based on symmetric correlations that are ubiquitous in resting-state functional MRI (fMRI). We use spectral dynamic causal modeling (DCM) to invert large graphs comprising dozens of nodes or regions. The ensuing graphs are directed and weighted, hence providing a neurobiologically plausible characterization of connectivity in terms of excitatory and inhibitory coupling. Furthermore, we show that the use of to discover the most likely sparse graph (or model) from a parent (e.g., fully connected) graph eschews the arbitrary thresholding often applied to large symmetric (functional connectivity) graphs. Using empirical fMRI data, we show that spectral DCM furnishes connectivity estimates on large graphs that correlate strongly with the estimates provided by stochastic DCM. Furthermore, we increase the efficiency of model inversion using functional connectivity modes to place prior constraints on effective connectivity. In other words, we use a small number of modes to finesse the potentially redundant parameterization of large DCMs. We show that spectral DCM-with functional connectivity priors-is ideally suited for directed graph theoretic analyses of resting-state fMRI. We envision that directed graphs will prove useful in understanding the psychopathology and pathophysiology of neurodegenerative and neurodevelopmental disorders. We will demonstrate the utility of large directed graphs in clinical populations in subsequent reports, using the procedures described in this paper.

  15. Searching for high-energy gamma-ray counterparts to gravitational-wave sources with Fermi-LAT: A needle in a haystack

    DOE PAGES

    Vianello, G.; Omodei, N.; Chiang, J.; ...

    2017-05-20

    At least a fraction of gravitational-wave (GW) progenitors are expected to emit an electromagnetic (EM) signal in the form of a short gamma-ray burst (sGRB). Discovering such a transient EM counterpart is challenging because the LIGO/VIRGO localization region is much larger (several hundreds of square degrees) than the field of view of X-ray, optical, and radio telescopes. The Fermi Large Area Telescope (LAT) has a wide field of view (~2.4 sr) and detects ~2–3 sGRBs per year above 100 MeV. It can detect them not only during the short prompt phase, but also during their long-lasting high-energy afterglow phase. If other wide-field, high-energy instruments such as Fermi-GBM, Swift-BAT, or INTEGRAL-ISGRI cannot detect or localize with enough precision an EM counterpart during the prompt phase, the LAT can potentially pinpoint it withmore » $$\\lesssim 10$$ arcmin accuracy during the afterglow phase. This routinely happens with gamma-ray bursts. Moreover, the LAT will cover the entire localization region within hours of any triggers during normal operations, allowing the γ-ray flux of any EM counterpart to be measured or constrained. As a result, we illustrate two new ad hoc methods to search for EM counterparts with the LAT and their application to the GW candidate LVT151012.« less

  16. Current and Near-Term Future Measurements of the Orbital Debris Environment at NASA

    NASA Technical Reports Server (NTRS)

    Stansbery, Gene; Liou, J.-C.; Mulrooney, M.; Horstman, M

    2010-01-01

    The NASA Orbital Debris Program Office places great emphasis on obtaining and understanding direct measurements of the orbital debris environment. The Orbital Debris Program Office's environmental models are all based on these measurements. Because OD measurements must cover a very wide range of sizes and altitudes, one technique realistically cannot be used for all measurements. In general, radar measurements have been used for lower altitudes and optical measurements for higher altitude orbits. For very small debris, in situ measurements such as returned spacecraft surfaces are utilized. In addition to receiving information from large debris (> 5-10 cm diameter) from the U.S. Space Surveillance Network, NASA conducts statistical measurements of the debris population for smaller sizes. NASA collects data from the Haystack and Goldstone radars for debris in low Earth orbit as small as 2- 4 mm diameter and from the Michigan Orbital DEbris Survey Telescope for debris near geosynchronous orbit altitude for sizes as small as 30-60 cm diameter. NASA is also currently examining the radiator panel of the Hubble Space Telescope Wide Field Planetary Camera 2 which was exposed to space for 16 years and was recently returned to Earth during the STS- 125 Space Shuttle mission. This paper will give an overview of these on-going measurement programs at NASA as well as discuss progress and plans for new instruments and techniques in the near future.

  17. Searching for high-energy gamma-ray counterparts to gravitational-wave sources with Fermi-LAT: A needle in a haystack

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Vianello, G.; Omodei, N.; Chiang, J.

    At least a fraction of gravitational-wave (GW) progenitors are expected to emit an electromagnetic (EM) signal in the form of a short gamma-ray burst (sGRB). Discovering such a transient EM counterpart is challenging because the LIGO/VIRGO localization region is much larger (several hundreds of square degrees) than the field of view of X-ray, optical, and radio telescopes. The Fermi Large Area Telescope (LAT) has a wide field of view (~2.4 sr) and detects ~2–3 sGRBs per year above 100 MeV. It can detect them not only during the short prompt phase, but also during their long-lasting high-energy afterglow phase. If other wide-field, high-energy instruments such as Fermi-GBM, Swift-BAT, or INTEGRAL-ISGRI cannot detect or localize with enough precision an EM counterpart during the prompt phase, the LAT can potentially pinpoint it withmore » $$\\lesssim 10$$ arcmin accuracy during the afterglow phase. This routinely happens with gamma-ray bursts. Moreover, the LAT will cover the entire localization region within hours of any triggers during normal operations, allowing the γ-ray flux of any EM counterpart to be measured or constrained. As a result, we illustrate two new ad hoc methods to search for EM counterparts with the LAT and their application to the GW candidate LVT151012.« less

  18. Preparing soft-bodied arthropods for microscope examination: Soft Scales (Insecta: Hemiptera: Coccidae)

    USDA-ARS?s Scientific Manuscript database

    Proper identification of soft scales (Hemiptera:Coccidae) requires preparation of the specimen on a microscope slide. This training video provides visual instruction on how to prepare soft scale specimens on microscope slides for examination and identification. Steps ranging from collection, speci...

  19. K-State Problem Identification Rating Scales for College Students

    ERIC Educational Resources Information Center

    Robertson, John M.; Benton, Stephen L.; Newton, Fred B.; Downey, Ronald G.; Marsh, Patricia A.; Benton, Sheryl A.; Tseng, Wen-Chih; Shin, Kang-Hyun

    2006-01-01

    The K-State Problem Identification Rating Scales, a new screening instrument for college counseling centers, gathers information about clients' presenting symptoms, functioning levels, and readiness to change. Three studies revealed 7 scales: Mood Difficulties, Learning Problems, Food Concerns, Interpersonal Conflicts, Career Uncertainties,…

  20. Prior knowledge based mining functional modules from Yeast PPI networks with gene ontology

    PubMed Central

    2010-01-01

    Background In the literature, there are fruitful algorithmic approaches for identification functional modules in protein-protein interactions (PPI) networks. Because of accumulation of large-scale interaction data on multiple organisms and non-recording interaction data in the existing PPI database, it is still emergent to design novel computational techniques that can be able to correctly and scalably analyze interaction data sets. Indeed there are a number of large scale biological data sets providing indirect evidence for protein-protein interaction relationships. Results The main aim of this paper is to present a prior knowledge based mining strategy to identify functional modules from PPI networks with the aid of Gene Ontology. Higher similarity value in Gene Ontology means that two gene products are more functionally related to each other, so it is better to group such gene products into one functional module. We study (i) to encode the functional pairs into the existing PPI networks; and (ii) to use these functional pairs as pairwise constraints to supervise the existing functional module identification algorithms. Topology-based modularity metric and complex annotation in MIPs will be used to evaluate the identified functional modules by these two approaches. Conclusions The experimental results on Yeast PPI networks and GO have shown that the prior knowledge based learning methods perform better than the existing algorithms. PMID:21172053

  1. Preparing soft-bodied arthropods for arthropods for microscope examination: Armored Scales (Insects: Hemiptera: Diaspididae)

    USDA-ARS?s Scientific Manuscript database

    Proper identification of armored scales (Hemiptera: Diaspididae) requires preparation of the specimen on a microscope slide. This training video provides visual instruction on how to prepare armored scales specimens on microscope slides for examination and identification. Steps ranging from collect...

  2. Damage identification of a TLP floating wind turbine by meta-heuristic algorithms

    NASA Astrophysics Data System (ADS)

    Ettefagh, M. M.

    2015-12-01

    Damage identification of the offshore floating wind turbine by vibration/dynamic signals is one of the important and new research fields in the Structural Health Monitoring (SHM). In this paper a new damage identification method is proposed based on meta-heuristic algorithms using the dynamic response of the TLP (Tension-Leg Platform) floating wind turbine structure. The Genetic Algorithms (GA), Artificial Immune System (AIS), Particle Swarm Optimization (PSO), and Artificial Bee Colony (ABC) are chosen for minimizing the object function, defined properly for damage identification purpose. In addition to studying the capability of mentioned algorithms in correctly identifying the damage, the effect of the response type on the results of identification is studied. Also, the results of proposed damage identification are investigated with considering possible uncertainties of the structure. Finally, for evaluating the proposed method in real condition, a 1/100 scaled experimental setup of TLP Floating Wind Turbine (TLPFWT) is provided in a laboratory scale and the proposed damage identification method is applied to the scaled turbine.

  3. Scale Insects, edition 2, a tool for the identification of potential pest scales at U.S.A. ports-of-entry (Hemiptera, Sternorrhyncha, Coccoidea)

    PubMed Central

    Miller, Douglass R.; Rung, Alessandra; Parikh, Grishma

    2014-01-01

    Abstract We provide a general overview of features and technical specifications of an online, interactive tool for the identification of scale insects of concern to the U.S.A. ports-of-entry. Full lists of terminal taxa included in the keys (of which there are four), a list of features used in them, and a discussion of the structure of the tool are provided. We also briefly discuss the advantages of interactive keys for the identification of potential scale insect pests. The interactive key is freely accessible on http://idtools.org/id/scales/index.php PMID:25152668

  4. Laboratory Needs for Interstellar Ice Studies

    NASA Astrophysics Data System (ADS)

    Boogert, Abraham C. A.

    2012-05-01

    A large fraction of the molecules in dense interstellar and circumstellar environments is stored in icy grain mantles. The mantles are formed by a complex interplay between chemical and physical processes. Key questions on the accretion and desorption processes and the chemistry on the grain surfaces and within the icy mantles can only be answered by laboratory experiments. Recent infrared (2-30 micron) spectroscopic surveys of large samples of Young Stellar Objects (YSOs) and background stars tracing quiescent cloud material have shown that the ice band profiles and depths vary considerably as a function of environment. Using laboratory spectra in the identification process, it is clear that a rather complex mixture of simple species (CH3OH, CO2, H2O, CO) exists even in the quiescent cloud phase. Variations of the local physical conditions (CO freeze out) and time scales (CH3OH formation) appear to be key factors in the observed variations. Sublimation and thermal processing dominate as YSOs heat their environments. The identification of several ice absorption features is still disputed. I will outline laboratory work (e.g., on salts, PAHs, and aliphatic hydrocarbons) needed to further constrain the ice band identification as well as the thermal and chemical history of the carriers. Such experiments will also be essential to interpret future high spectral resolution SOFIA and JWST observations.

  5. A regulation probability model-based meta-analysis of multiple transcriptomics data sets for cancer biomarker identification.

    PubMed

    Xie, Xin-Ping; Xie, Yu-Feng; Wang, Hong-Qiang

    2017-08-23

    Large-scale accumulation of omics data poses a pressing challenge of integrative analysis of multiple data sets in bioinformatics. An open question of such integrative analysis is how to pinpoint consistent but subtle gene activity patterns across studies. Study heterogeneity needs to be addressed carefully for this goal. This paper proposes a regulation probability model-based meta-analysis, jGRP, for identifying differentially expressed genes (DEGs). The method integrates multiple transcriptomics data sets in a gene regulatory space instead of in a gene expression space, which makes it easy to capture and manage data heterogeneity across studies from different laboratories or platforms. Specifically, we transform gene expression profiles into a united gene regulation profile across studies by mathematically defining two gene regulation events between two conditions and estimating their occurring probabilities in a sample. Finally, a novel differential expression statistic is established based on the gene regulation profiles, realizing accurate and flexible identification of DEGs in gene regulation space. We evaluated the proposed method on simulation data and real-world cancer datasets and showed the effectiveness and efficiency of jGRP in identifying DEGs identification in the context of meta-analysis. Data heterogeneity largely influences the performance of meta-analysis of DEGs identification. Existing different meta-analysis methods were revealed to exhibit very different degrees of sensitivity to study heterogeneity. The proposed method, jGRP, can be a standalone tool due to its united framework and controllable way to deal with study heterogeneity.

  6. Basin-scale heterogeneity in Antarctic precipitation and its impact on surface mass variability

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fyke, Jeremy; Lenaerts, Jan T. M.; Wang, Hailong

    Annually averaged precipitation in the form of snow, the dominant term of the Antarctic Ice Sheet surface mass balance, displays large spatial and temporal variability. Here we present an analysis of spatial patterns of regional Antarctic precipitation variability and their impact on integrated Antarctic surface mass balance variability simulated as part of a preindustrial 1800-year global, fully coupled Community Earth System Model simulation. Correlation and composite analyses based on this output allow for a robust exploration of Antarctic precipitation variability. We identify statistically significant relationships between precipitation patterns across Antarctica that are corroborated by climate reanalyses, regional modeling and icemore » core records. These patterns are driven by variability in large-scale atmospheric moisture transport, which itself is characterized by decadal- to centennial-scale oscillations around the long-term mean. We suggest that this heterogeneity in Antarctic precipitation variability has a dampening effect on overall Antarctic surface mass balance variability, with implications for regulation of Antarctic-sourced sea level variability, detection of an emergent anthropogenic signal in Antarctic mass trends and identification of Antarctic mass loss accelerations.« less

  7. Basin-scale heterogeneity in Antarctic precipitation and its impact on surface mass variability

    DOE PAGES

    Fyke, Jeremy; Lenaerts, Jan T. M.; Wang, Hailong

    2017-11-15

    Annually averaged precipitation in the form of snow, the dominant term of the Antarctic Ice Sheet surface mass balance, displays large spatial and temporal variability. Here we present an analysis of spatial patterns of regional Antarctic precipitation variability and their impact on integrated Antarctic surface mass balance variability simulated as part of a preindustrial 1800-year global, fully coupled Community Earth System Model simulation. Correlation and composite analyses based on this output allow for a robust exploration of Antarctic precipitation variability. We identify statistically significant relationships between precipitation patterns across Antarctica that are corroborated by climate reanalyses, regional modeling and icemore » core records. These patterns are driven by variability in large-scale atmospheric moisture transport, which itself is characterized by decadal- to centennial-scale oscillations around the long-term mean. We suggest that this heterogeneity in Antarctic precipitation variability has a dampening effect on overall Antarctic surface mass balance variability, with implications for regulation of Antarctic-sourced sea level variability, detection of an emergent anthropogenic signal in Antarctic mass trends and identification of Antarctic mass loss accelerations.« less

  8. Finding My Needle in the Haystack: Effective Personalized Re-ranking of Search Results in Prospector

    NASA Astrophysics Data System (ADS)

    König, Florian; van Velsen, Lex; Paramythis, Alexandros

    This paper provides an overview of Prospector, a personalized Internet meta-search engine, which utilizes a combination of ontological information, ratings-based models of user interests, and complementary theme-oriented group models to recommend (through re-ranking) search results obtained from an underlying search engine. Re-ranking brings “closer to the top” those items that are of particular interest to a user or have high relevance to a given theme. A user-based, real-world evaluation has shown that the system is effective in promoting results of interest, but lags behind Google in user acceptance, possibly due to the absence of features popularized by said search engine. Overall, users would consider employing a personalized search engine to perform searches with terms that require disambiguation and / or contextualization.

  9. A Survey of School Psychologists' Practices for Identifying Mentally Retarded Students.

    ERIC Educational Resources Information Center

    Wodrich, David L.; Barry, Christine T.

    1991-01-01

    Surveyed school psychologists regarding identification of mentally retarded students. The Wechsler scales were the most frequently used tests for deriving intelligence quotient scores, which together with adaptive behavior scale scores were rated as most influential in identification-placement decisions. The Vineland Adaptive Behavior Scales were…

  10. Towards large-scale FAME-based bacterial species identification using machine learning techniques.

    PubMed

    Slabbinck, Bram; De Baets, Bernard; Dawyndt, Peter; De Vos, Paul

    2009-05-01

    In the last decade, bacterial taxonomy witnessed a huge expansion. The swift pace of bacterial species (re-)definitions has a serious impact on the accuracy and completeness of first-line identification methods. Consequently, back-end identification libraries need to be synchronized with the List of Prokaryotic names with Standing in Nomenclature. In this study, we focus on bacterial fatty acid methyl ester (FAME) profiling as a broadly used first-line identification method. From the BAME@LMG database, we have selected FAME profiles of individual strains belonging to the genera Bacillus, Paenibacillus and Pseudomonas. Only those profiles resulting from standard growth conditions have been retained. The corresponding data set covers 74, 44 and 95 validly published bacterial species, respectively, represented by 961, 378 and 1673 standard FAME profiles. Through the application of machine learning techniques in a supervised strategy, different computational models have been built for genus and species identification. Three techniques have been considered: artificial neural networks, random forests and support vector machines. Nearly perfect identification has been achieved at genus level. Notwithstanding the known limited discriminative power of FAME analysis for species identification, the computational models have resulted in good species identification results for the three genera. For Bacillus, Paenibacillus and Pseudomonas, random forests have resulted in sensitivity values, respectively, 0.847, 0.901 and 0.708. The random forests models outperform those of the other machine learning techniques. Moreover, our machine learning approach also outperformed the Sherlock MIS (MIDI Inc., Newark, DE, USA). These results show that machine learning proves very useful for FAME-based bacterial species identification. Besides good bacterial identification at species level, speed and ease of taxonomic synchronization are major advantages of this computational species identification strategy.

  11. Next Generation Astronomical Data Processing using Big Data Technologies from the Apache Software Foundation

    NASA Astrophysics Data System (ADS)

    Mattmann, Chris

    2014-04-01

    In this era of exascale instruments for astronomy we must naturally develop next generation capabilities for the unprecedented data volume and velocity that will arrive due to the veracity of these ground-based sensor and observatories. Integrating scientific algorithms stewarded by scientific groups unobtrusively and rapidly; intelligently selecting data movement technologies; making use of cloud computing for storage and processing; and automatically extracting text and metadata and science from any type of file are all needed capabilities in this exciting time. Our group at NASA JPL has promoted the use of open source data management technologies available from the Apache Software Foundation (ASF) in pursuit of constructing next generation data management and processing systems for astronomical instruments including the Expanded Very Large Array (EVLA) in Socorro, NM and the Atacama Large Milimetre/Sub Milimetre Array (ALMA); as well as for the KAT-7 project led by SKA South Africa as a precursor to the full MeerKAT telescope. In addition we are funded currently by the National Science Foundation in the US to work with MIT Haystack Observatory and the University of Cambridge in the UK to construct a Radio Array of Portable Interferometric Devices (RAPID) that will undoubtedly draw from the rich technology advances underway. NASA JPL is investing in a strategic initiative for Big Data that is pulling in these capabilities and technologies for astronomical instruments and also for Earth science remote sensing. In this talk I will describe the above collaborative efforts underway and point to solutions in open source from the Apache Software Foundation that can be deployed and used today and that are already bringing our teams and projects benefits. I will describe how others can take advantage of our experience and point towards future application and contribution of these tools.

  12. JHelioviewer: Open-Source Software for Discovery and Image Access in the Petabyte Age

    NASA Astrophysics Data System (ADS)

    Mueller, D.; Dimitoglou, G.; Garcia Ortiz, J.; Langenberg, M.; Nuhn, M.; Dau, A.; Pagel, S.; Schmidt, L.; Hughitt, V. K.; Ireland, J.; Fleck, B.

    2011-12-01

    The unprecedented torrent of data returned by the Solar Dynamics Observatory is both a blessing and a barrier: a blessing for making available data with significantly higher spatial and temporal resolution, but a barrier for scientists to access, browse and analyze them. With such staggering data volume, the data is accessible only from a few repositories and users have to deal with data sets effectively immobile and practically difficult to download. From a scientist's perspective this poses three challenges: accessing, browsing and finding interesting data while avoiding the proverbial search for a needle in a haystack. To address these challenges, we have developed JHelioviewer, an open-source visualization software that lets users browse large data volumes both as still images and movies. We did so by deploying an efficient image encoding, storage, and dissemination solution using the JPEG 2000 standard. This solution enables users to access remote images at different resolution levels as a single data stream. Users can view, manipulate, pan, zoom, and overlay JPEG 2000 compressed data quickly, without severe network bandwidth penalties. Besides viewing data, the browser provides third-party metadata and event catalog integration to quickly locate data of interest, as well as an interface to the Virtual Solar Observatory to download science-quality data. As part of the ESA/NASA Helioviewer Project, JHelioviewer offers intuitive ways to browse large amounts of heterogeneous data remotely and provides an extensible and customizable open-source platform for the scientific community. In addition, the easy-to-use graphical user interface enables the general public and educators to access, enjoy and reuse data from space missions without barriers.

  13. Machine Learning Approach to Automated Quality Identification of Human Induced Pluripotent Stem Cell Colony Images.

    PubMed

    Joutsijoki, Henry; Haponen, Markus; Rasku, Jyrki; Aalto-Setälä, Katriina; Juhola, Martti

    2016-01-01

    The focus of this research is on automated identification of the quality of human induced pluripotent stem cell (iPSC) colony images. iPS cell technology is a contemporary method by which the patient's cells are reprogrammed back to stem cells and are differentiated to any cell type wanted. iPS cell technology will be used in future to patient specific drug screening, disease modeling, and tissue repairing, for instance. However, there are technical challenges before iPS cell technology can be used in practice and one of them is quality control of growing iPSC colonies which is currently done manually but is unfeasible solution in large-scale cultures. The monitoring problem returns to image analysis and classification problem. In this paper, we tackle this problem using machine learning methods such as multiclass Support Vector Machines and several baseline methods together with Scaled Invariant Feature Transformation based features. We perform over 80 test arrangements and do a thorough parameter value search. The best accuracy (62.4%) for classification was obtained by using a k-NN classifier showing improved accuracy compared to earlier studies.

  14. Gram-scale purification of aconitine and identification of lappaconitine in Aconitum karacolicum.

    PubMed

    Tarbe, M; de Pomyers, H; Mugnier, L; Bertin, D; Ibragimov, T; Gigmes, D; Mabrouk, K

    2017-07-01

    Aconitum karacolicum from northern Kyrgyzstan (Alatau area) contains about 0.8-1% aconitine as well as other aconite derivatives that have already been identified. In this paper, we compare several methods for the further purification of an Aconitum karacolicum extract initially containing 80% of aconitine. Reverse-phase flash chromatography, reverse-phase semi-preparative HPLC, centrifugal partition chromatography (CPC) and recrystallization techniques were evaluated regarding first their efficiency to get the highest purity of aconitine (over 96%) and secondly their applicability in a semi-industrial scale purification process (in our case, 150g of plant extract). Even if the CPC technique shows the highest purification yield (63%), the recrystallization remains the method of choice to purify a large amount of aconitine as i) it can be easily carried out in safe conditions; ii) an aprotic solvent is used, avoiding aconitine degradation. Moreover, this study led us to the identification of lappaconitine in Aconitum karacolicum, a well-known alkaloid never found in this Aconitum species. Copyright © 2017 Elsevier B.V. All rights reserved.

  15. Genome resequencing in Populus: Revealing large-scale genome variation and implications on specialized-trait genomics

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Muchero, Wellington; Labbe, Jessy L; Priya, Ranjan

    2014-01-01

    To date, Populus ranks among a few plant species with a complete genome sequence and other highly developed genomic resources. With the first genome sequence among all tree species, Populus has been adopted as a suitable model organism for genomic studies in trees. However, far from being just a model species, Populus is a key renewable economic resource that plays a significant role in providing raw materials for the biofuel and pulp and paper industries. Therefore, aside from leading frontiers of basic tree molecular biology and ecological research, Populus leads frontiers in addressing global economic challenges related to fuel andmore » fiber production. The latter fact suggests that research aimed at improving quality and quantity of Populus as a raw material will likely drive the pursuit of more targeted and deeper research in order to unlock the economic potential tied in molecular biology processes that drive this tree species. Advances in genome sequence-driven technologies, such as resequencing individual genotypes, which in turn facilitates large scale SNP discovery and identification of large scale polymorphisms are key determinants of future success in these initiatives. In this treatise we discuss implications of genome sequence-enable technologies on Populus genomic and genetic studies of complex and specialized-traits.« less

  16. Potential slab avalanche release area identification from estimated winter terrain: a multi-scale, fuzzy logic approach

    NASA Astrophysics Data System (ADS)

    Veitinger, Jochen; Purves, Ross Stuart; Sovilla, Betty

    2016-10-01

    Avalanche hazard assessment requires a very precise estimation of the release area, which still depends, to a large extent, on expert judgement of avalanche specialists. Therefore, a new algorithm for automated identification of potential avalanche release areas was developed. It overcomes some of the limitations of previous tools, which are currently not often applied in hazard mitigation practice. By introducing a multi-scale roughness parameter, fine-scale topography and its attenuation under snow influence is captured. This allows the assessment of snow influence on terrain morphology and, consequently, potential release area size and location. The integration of a wind shelter index enables the user to define release area scenarios as a function of the prevailing wind direction or single storm events. A case study illustrates the practical usefulness of this approach for the definition of release area scenarios under varying snow cover and wind conditions. A validation with historical data demonstrated an improved estimation of avalanche release areas. Our method outperforms a slope-based approach, in particular for more frequent avalanches; however, the application of the algorithm as a forecasting tool remains limited, as snowpack stability is not integrated. Future research activity should therefore focus on the coupling of the algorithm with snowpack conditions.

  17. Metabolic profiling and systematic identification of flavonoids and isoflavonoids in roots and cell suspension cultures of Medicago truncatula using HPLC-UV-ESI-MS and GC-MS.

    PubMed

    Farag, Mohamed A; Huhman, David V; Lei, Zhentian; Sumner, Lloyd W

    2007-02-01

    An integrated approach utilizing HPLC-UV-ESI-MS and GC-MS was used for the large-scale and systematic identification of polyphenols in Medicago truncatula root and cell culture. Under optimized conditions, we were able to simultaneously quantify and identify 35 polyphenols including 26 isoflavones, 3 flavones, 2 flavanones, 2 aurones and a chalcone. All identifications were based upon UV spectra, mass spectral characteristics of protonated molecules, tandem mass spectral data, mass measurements obtained using a quadrupole time-of-flight mass spectrometer (QtofMS), and confirmed through the co-characterization of authentic compounds. In specific instances where the stereochemistry of sugar conjugates was uncertain, subsequent enzymatic hydrolysis of the conjugate followed by GC-MS was used to assign the sugar stereochemical configuration. Comparative metabolic profiling of Medicago truncatula root and cell cultures was then performed and revealed significant differences in the isoflavonoid composition of these two tissues.

  18. Detection, breakpoint identification and detailed characterisation of a CNV at the FRA16D site using SNP assays.

    PubMed

    Winchester, L; Newbury, D F; Monaco, A P; Ragoussis, J

    2008-01-01

    Copy Number Variants (CNV) and other submicroscopic structural changes are now recognised to be widespread across the human genome. We show that SNP data generated for association study can be utilised for the identification of deletion CNVs. During analysis of data for an SNP association study for Specific Language Impairment (SLI) a deletion was identified. SLI adversely affects the language development of children in the absence of any obvious cause. Previous studies have found linkage to a region on chromosome 16. The deletion was located in a known fragile site FRA16D in intron 5-6 of the WWOX gene (also known as FOR). Changes in the FRA16D site have been previously linked to cancer and are often characterised in cell lines. A long-range PCR assay was used to confirm the existence of the deletion. We also show the breakpoint identification and large-scale characterisation of this CNV in a normal human sample set. Copyright 2009 S. Karger AG, Basel.

  19. Technological advancements and their importance for nematode identification

    NASA Astrophysics Data System (ADS)

    Ahmed, Mohammed; Sapp, Melanie; Prior, Thomas; Karssen, Gerrit; Back, Matthew Alan

    2016-06-01

    Nematodes represent a species-rich and morphologically diverse group of metazoans known to inhabit both aquatic and terrestrial environments. Their role as biological indicators and as key players in nutrient cycling has been well documented. Some plant-parasitic species are also known to cause significant losses to crop production. In spite of this, there still exists a huge gap in our knowledge of their diversity due to the enormity of time and expertise often involved in characterising species using phenotypic features. Molecular methodology provides useful means of complementing the limited number of reliable diagnostic characters available for morphology-based identification. We discuss herein some of the limitations of traditional taxonomy and how molecular methodologies, especially the use of high-throughput sequencing, have assisted in carrying out large-scale nematode community studies and characterisation of phytonematodes through rapid identification of multiple taxa. We also provide brief descriptions of some the current and almost-outdated high-throughput sequencing platforms and their applications in both plant nematology and soil ecology.

  20. Method for identification of rigid domains and hinge residues in proteins based on exhaustive enumeration.

    PubMed

    Sim, Jaehyun; Sim, Jun; Park, Eunsung; Lee, Julian

    2015-06-01

    Many proteins undergo large-scale motions where relatively rigid domains move against each other. The identification of rigid domains, as well as the hinge residues important for their relative movements, is important for various applications including flexible docking simulations. In this work, we develop a method for protein rigid domain identification based on an exhaustive enumeration of maximal rigid domains, the rigid domains not fully contained within other domains. The computation is performed by mapping the problem to that of finding maximal cliques in a graph. A minimal set of rigid domains are then selected, which cover most of the protein with minimal overlap. In contrast to the results of existing methods that partition a protein into non-overlapping domains using approximate algorithms, the rigid domains obtained from exact enumeration naturally contain overlapping regions, which correspond to the hinges of the inter-domain bending motion. The performance of the algorithm is demonstrated on several proteins. © 2015 Wiley Periodicals, Inc.

  1. [Identification of mouse brain neuropeptides by high throughput mass spectrometry].

    PubMed

    Shao, Xianfeng; Ma, Min; Chen, Ruibing; Jia, Chenxi

    2018-04-25

    Neuropeptides play an important role in the physiological functions of the human body. The physiological activities such as pain, sleep, mood, learning and memory are affected by neuropeptides. Neuropeptides mainly exist in the nerve tissue of the body, and a small amount of them are distributed in body fluid and organs. At present, analysis of large-scale identification of neuropeptides in whole brain tissue is still challenging. Therefore, high-throughput detection of these neuropeptides is greatly significant to understand the composition and function of neuropeptides. In this study, 1 830 endogenous peptides and 99 novel putative neuropeptides were identified by extraction of endogenous peptides from whole brain tissue of mice by liquid phase tandem mass spectrometry (LC-MS / MS). The identification of these endogenous peptides provides not only a reference value in the treatment and mechanism studies of diseases and the development of drugs, but also the basis for the study of a new neuropeptides and their functions.

  2. Detection of High Energy Cosmic Ray with the Advanced Thin Ionization Calorimeter (ATIC)

    NASA Technical Reports Server (NTRS)

    Fazely, Ali R.

    2003-01-01

    ATIC is a balloon-borne investigation of cosmic ray spectra, from below 50 GeV to near 100 TeV total energy, using a fully active Bismuth Gemmate (BGO) calorimeter. It is equipped with the first large area mosaic of small fully depleted silicon detector pixels capable of charge identification in cosmic rays from H to Fe. As a redundancy check for the charge identification and a coarse particle tracking system, three projective layers of x-y scintillator hodoscopes were employed, above, in the center and below a Carbon interaction 'target'. Very high energy gamma-rays and their energy spectrum may provide insight to the flux of extremely high energy neutrinos which will be investigated in detail with several proposed cubic kilometer scale neutrino observatories in the next decade.

  3. VIP Barcoding: composition vector-based software for rapid species identification based on DNA barcoding.

    PubMed

    Fan, Long; Hui, Jerome H L; Yu, Zu Guo; Chu, Ka Hou

    2014-07-01

    Species identification based on short sequences of DNA markers, that is, DNA barcoding, has emerged as an integral part of modern taxonomy. However, software for the analysis of large and multilocus barcoding data sets is scarce. The Basic Local Alignment Search Tool (BLAST) is currently the fastest tool capable of handling large databases (e.g. >5000 sequences), but its accuracy is a concern and has been criticized for its local optimization. However, current more accurate software requires sequence alignment or complex calculations, which are time-consuming when dealing with large data sets during data preprocessing or during the search stage. Therefore, it is imperative to develop a practical program for both accurate and scalable species identification for DNA barcoding. In this context, we present VIP Barcoding: a user-friendly software in graphical user interface for rapid DNA barcoding. It adopts a hybrid, two-stage algorithm. First, an alignment-free composition vector (CV) method is utilized to reduce searching space by screening a reference database. The alignment-based K2P distance nearest-neighbour method is then employed to analyse the smaller data set generated in the first stage. In comparison with other software, we demonstrate that VIP Barcoding has (i) higher accuracy than Blastn and several alignment-free methods and (ii) higher scalability than alignment-based distance methods and character-based methods. These results suggest that this platform is able to deal with both large-scale and multilocus barcoding data with accuracy and can contribute to DNA barcoding for modern taxonomy. VIP Barcoding is free and available at http://msl.sls.cuhk.edu.hk/vipbarcoding/. © 2014 John Wiley & Sons Ltd.

  4. A possible approach to large-scale laboratory testing for acute radiation sickness after a nuclear detonation.

    PubMed

    Adalja, Amesh A; Watson, Matthew; Wollner, Samuel; Toner, Eric

    2011-12-01

    After the detonation of an improvised nuclear device, several key actions will be necessary to save the greatest number of lives possible. Among these tasks, the identification of patients with impending acute radiation sickness is a critical problem that so far has lacked a clear solution in national planning. We present one possible solution: the formation of a public-private partnership to augment the capacity to identify those at risk for acute radiation sickness. © Mary Ann Liebert, Inc.

  5. Large-Scale Interaction Profiling of Protein Domains Through Proteomic Peptide-Phage Display Using Custom Peptidomes.

    PubMed

    Seo, Moon-Hyeong; Nim, Satra; Jeon, Jouhyun; Kim, Philip M

    2017-01-01

    Protein-protein interactions are essential to cellular functions and signaling pathways. We recently combined bioinformatics and custom oligonucleotide arrays to construct custom-made peptide-phage libraries for screening peptide-protein interactions, an approach we call proteomic peptide-phage display (ProP-PD). In this chapter, we describe protocols for phage display for the identification of natural peptide binders for a given protein. We finally describe deep sequencing for the analysis of the proteomic peptide-phage display.

  6. The OncoPPi network of cancer-focused protein-protein interactions to inform biological insights and therapeutic strategies* | Office of Cancer Genomics

    Cancer.gov

    As genomics advances reveal the cancer gene landscape, a daunting task is to understand how these genes contribute to dysregulated oncogenic pathways. Integration of cancer genes into networks offers opportunities to reveal protein–protein interactions (PPIs) with functional and therapeutic significance. Here, we report the generation of a cancer-focused PPI network, termed OncoPPi, and identification of >260 cancer-associated PPIs not in other large-scale interactomes.

  7. [The prospect of application of toxicogenetics/pharmcogenetics theory and methods in forensic practice].

    PubMed

    Shen, Dan-na; Yi, Xu-fu; Chen, Xiao-gang; Xu, Tong-li; Cui, Li-juan

    2007-10-01

    Individual response to drugs, toxicants, environmental chemicals and allergens varies with genotype. Some respond well to these substances without significant consequences, while others may respond strongly with severe consequences and even death. Toxicogenetics and toxicogenomics as well as pharmacogenetics explain the genetic basis for the variations of individual response to toxicants by sequencing the human genome and large-scale identification of genome polymorphism. The new disciplines will provide a new route for forensic specialists to determine the cause of death.

  8. International Workshop on Structural and Functional Aspects of the Cholinergic Synapse Held in Jerusalem, Isreal on 30 August-4 September 1987

    DTIC Science & Technology

    1987-09-01

    77) Large scale purification of the acetylcholine receptor protein In its membrane-bound and detergent extracted forms from Torpedo marmorata...maintenance of the postsynaptic apparatus in the adult. Our studies have alac led to the Identification of agrin, a protein that is extracted from the synapse...in extracts of muscle, and monoclonal antibodies directed against &grin recognize molecules highly concentrated in the synaptic basal lamina at the

  9. Geologic Reconnaissance and Lithologic Identification by Remote Sensing

    DTIC Science & Technology

    remote sensing in geologic reconnaissance for purposes of tunnel site selection was studied further and a test case was undertaken to evaluate this geological application. Airborne multispectral scanning (MSS) data were obtained in May, 1972, over a region between Spearfish and Rapid City, South Dakota. With major effort directed toward the analysis of these data, the following geologic features were discriminated: (1) exposed rock areas, (2) five separate rock groups, (3) large-scale structures. This discrimination was accomplished by ratioing multispectral channels.

  10. Manual of downburst identification for Project NIMROD. [atmospheric circulation

    NASA Technical Reports Server (NTRS)

    Fujita, T. T.

    1978-01-01

    Aerial photography, Doppler radar, and satellite infrared imagery are used in the two year National Intensive Meteorological Research on Downburst (NIMROD) project to provide large area mapping of strong downdrafts that induce an outward burst of damaging winds over or near the earth. Topics discussed include scales of thunderstorm outflow; aerial photographs of downburst damage; microbursts and aviation hazards; radar echo characteristics; infrared imagery from GOES/SMS; and downburts-tornado relationships. Color maps of downbursts and tornadoes are included.

  11. Relative scale and the strength and deformability of rock masses

    NASA Astrophysics Data System (ADS)

    Schultz, Richard A.

    1996-09-01

    The strength and deformation of rocks depend strongly on the degree of fracturing, which can be assessed in the field and related systematically to these properties. Appropriate Mohr envelopes obtained from the Rock Mass Rating (RMR) classification system and the Hoek-Brown criterion for outcrops and other large-scale exposures of fractured rocks show that rock-mass cohesive strength, tensile strength, and unconfined compressive strength can be reduced by as much as a factor often relative to values for the unfractured material. The rock-mass deformation modulus is also reduced relative to Young's modulus. A "cook-book" example illustrates the use of RMR in field applications. The smaller values of rock-mass strength and deformability imply that there is a particular scale of observation whose identification is critical to applying laboratory measurements and associated failure criteria to geologic structures.

  12. Religion and National Identification in Europe: Comparing Muslim Youth in Belgium, England, Germany, the Netherlands, and Sweden

    PubMed Central

    Fleischmann, Fenella; Phalet, Karen

    2017-01-01

    How inclusive are European national identities of Muslim minorities and how can we explain cross-cultural variation in inclusiveness? To address these questions, we draw on large-scale school-based surveys of Muslim minority and non-Muslim majority and other minority youth in five European countries (Children of Immigrants Longitudinal Survey [CILS]; Belgium, England, Germany, the Netherlands, and Sweden). Our double comparison of national identification across groups and countries reveals that national identities are less strongly endorsed by all minorities compared with majority youth, but national identification is lowest among Muslims. This descriptive evidence resonates with public concerns about the insufficient inclusion of immigrant minorities in general, and Muslims in particular, in European national identities. In addition, significant country variation in group differences in identification suggest that some national identities are more inclusive of Muslims than others. Taking an intergroup relations approach to the inclusiveness of national identities for Muslims, we establish that beyond religious commitment, positive intergroup contact (majority friendship) plays a major role in explaining differences in national identification in multigroup multilevel mediation models, whereas experiences of discrimination in school do not contribute to this explanation. Our comparative findings thus establish contextual variation in the inclusiveness of intergroup relations and European national identities for Muslim minorities. PMID:29386688

  13. Maximizing the sensitivity and reliability of peptide identification in large-scale proteomic experiments by harnessing multiple search engines.

    PubMed

    Yu, Wen; Taylor, J Alex; Davis, Michael T; Bonilla, Leo E; Lee, Kimberly A; Auger, Paul L; Farnsworth, Chris C; Welcher, Andrew A; Patterson, Scott D

    2010-03-01

    Despite recent advances in qualitative proteomics, the automatic identification of peptides with optimal sensitivity and accuracy remains a difficult goal. To address this deficiency, a novel algorithm, Multiple Search Engines, Normalization and Consensus is described. The method employs six search engines and a re-scoring engine to search MS/MS spectra against protein and decoy sequences. After the peptide hits from each engine are normalized to error rates estimated from the decoy hits, peptide assignments are then deduced using a minimum consensus model. These assignments are produced in a series of progressively relaxed false-discovery rates, thus enabling a comprehensive interpretation of the data set. Additionally, the estimated false-discovery rate was found to have good concordance with the observed false-positive rate calculated from known identities. Benchmarking against standard proteins data sets (ISBv1, sPRG2006) and their published analysis, demonstrated that the Multiple Search Engines, Normalization and Consensus algorithm consistently achieved significantly higher sensitivity in peptide identifications, which led to increased or more robust protein identifications in all data sets compared with prior methods. The sensitivity and the false-positive rate of peptide identification exhibit an inverse-proportional and linear relationship with the number of participating search engines.

  14. Identification of varying time scales in sediment transport using the Hilbert-Huang Transform method

    NASA Astrophysics Data System (ADS)

    Kuai, Ken Z.; Tsai, Christina W.

    2012-02-01

    SummarySediment transport processes vary at a variety of time scales - from seconds, hours, days to months and years. Multiple time scales exist in the system of flow, sediment transport and bed elevation change processes. As such, identification and selection of appropriate time scales for flow and sediment processes can assist in formulating a system of flow and sediment governing equations representative of the dynamic interaction of flow and particles at the desired details. Recognizing the importance of different varying time scales in the fluvial processes of sediment transport, we introduce the Hilbert-Huang Transform method (HHT) to the field of sediment transport for the time scale analysis. The HHT uses the Empirical Mode Decomposition (EMD) method to decompose a time series into a collection of the Intrinsic Mode Functions (IMFs), and uses the Hilbert Spectral Analysis (HSA) to obtain instantaneous frequency data. The EMD extracts the variability of data with different time scales, and improves the analysis of data series. The HSA can display the succession of time varying time scales, which cannot be captured by the often-used Fast Fourier Transform (FFT) method. This study is one of the earlier attempts to introduce the state-of-the-art technique for the multiple time sales analysis of sediment transport processes. Three practical applications of the HHT method for data analysis of both suspended sediment and bedload transport time series are presented. The analysis results show the strong impact of flood waves on the variations of flow and sediment time scales at a large sampling time scale, as well as the impact of flow turbulence on those time scales at a smaller sampling time scale. Our analysis reveals that the existence of multiple time scales in sediment transport processes may be attributed to the fractal nature in sediment transport. It can be demonstrated by the HHT analysis that the bedload motion time scale is better represented by the ratio of the water depth to the settling velocity, h/ w. In the final part, HHT results are compared with an available time scale formula in literature.

  15. Predicting Regional Self-identification from Spatial Network Models

    PubMed Central

    Almquist, Zack W.; Butts, Carter T.

    2014-01-01

    Social scientists characterize social life as a hierarchy of environments, from the micro level of an individual’s knowledge and perceptions to the macro level of large-scale social networks. In accordance with this typology, individuals are typically thought to reside in micro- and macro-level structures, composed of multifaceted relations (e.g., acquaintanceship, friendship, and kinship). This article analyzes the effects of social structure on micro outcomes through the case of regional identification. Self identification occurs in many different domains, one of which is regional; i.e., the identification of oneself with a locationally-associated group (e.g., a “New Yorker” or “Parisian”). Here, regional self-identification is posited to result from an influence process based on the location of an individual’s alters (e.g., friends, kin or coworkers), such that one tends to identify with regions in which many of his or her alters reside. The structure of this paper is laid out as follows: initially, we begin with a discussion of the relevant social science literature for both social networks and identification. This discussion is followed with one about competing mechanisms for regional identification that are motivated first from the social network literature, and second by the social psychological and cognitive literature of decision making and heuristics. Next, the paper covers the data and methods employed to test the proposed mechanisms. Finally, the paper concludes with a discussion of its findings and further implications for the larger social science literature. PMID:25684791

  16. Finite element modeling and analysis of tires

    NASA Technical Reports Server (NTRS)

    Noor, A. K.; Andersen, C. M.

    1983-01-01

    Predicting the response of tires under various loading conditions using finite element technology is addressed. Some of the recent advances in finite element technology which have high potential for application to tire modeling problems are reviewed. The analysis and modeling needs for tires are identified. Reduction methods for large-scale nonlinear analysis, with particular emphasis on treatment of combined loads, displacement-dependent and nonconservative loadings; development of simple and efficient mixed finite element models for shell analysis, identification of equivalent mixed and purely displacement models, and determination of the advantages of using mixed models; and effective computational models for large-rotation nonlinear problems, based on a total Lagrangian description of the deformation are included.

  17. Integrating Green and Blue Water Management Tools for Land and Water Resources Planning

    NASA Astrophysics Data System (ADS)

    Jewitt, G. P. W.

    2009-04-01

    The role of land use and land use change on the hydrological cycle is well known. However, the impacts of large scale land use change are poorly considered in water resources planning, unless they require direct abstraction of water resources and associated development of infrastructure e.g. Irrigation Schemes. However, large scale deforestation for the supply of raw materials, expansion of the areas of plantation forestry, increasing areas under food production and major plans for cultivation of biofuels in many developing countries are likely to result in extensive land use change. Given the spatial extent and temporal longevity of these proposed developments, major impacts on water resources are inevitable. It is imperative that managers and planners consider the consequences for downstream ecosystems and users in such developments. However, many popular tools, such as the vitual water approach, provide only coarse scale "order of magnitude" type estimates with poor consideration of, and limited usefulness, for land use planning. In this paper, a framework for the consideration of the impacts of large scale land use change on water resources at a range of temporal and spatial scales is presented. Drawing on experiences from South Africa, where the establishment of exotic commercial forest plantations is only permitted once a water use license has been granted, the framework adopts the "green water concept" for the identification of potential high impact areas of land use change and provides for integration with traditional "blue water" water resources planning tools for more detailed planning. Appropriate tools, ranging from simple spreadsheet solutions to more sophisticated remote sensing and hydrological models are described, and the application of the framework for consideration of water resources impacts associated with the establishment of large scale tectona grandis, sugar cane and jatropha curcas plantations is illustrated through examples in Mozambique and South Africa. Keywords: Land use change, water resources, green water, blue water, biofuels, developing countries

  18. Effects of microhabitat and land use on stream salamander abundance in the southwest Virginia coalfields

    USGS Publications Warehouse

    Sweeten, Sara E.; Ford, W. Mark

    2015-01-01

    Large-scale land uses such as residential wastewater discharge and coal mining practices, particularly surface coal extraction and associated valley fills, are of particular ecological concern in central Appalachia. Identification and quantification of both alterations across scales are a necessary first-step to mitigate negative consequences to biota. In central Appalachian headwater streams absent of fish, salamanders are the dominant, most abundant vertebrate predator providing a significant intermediate trophic role. Stream salamander species are considered to be sensitive to aquatic stressors and environmental alterations, and past research has shown linkages among microhabitat parameters, large-scale land use such as urbanization and logging with salamander abundances. However, little is known about these linkages in the coalfields of central Appalachia. In the summer of 2013, we visited 70 sites (sampled three times each) in the southwest Virginia coalfields to survey salamanders and quantify stream and riparian microhabitat parameters. Using an information-theoretic framework we compared the effects of microhabitat and large-scale land use on salamander abundances. Our findings indicate that dusky salamander (Desmognathus spp.) abundances are more correlated to microhabitat parameters such as canopy cover than to subwatershed land uses. Brook salamander (Eurycea spp.) abundances show strong negative associations to the suspended sediments and stream substrate embeddedness. Neither Desmognathus spp. nor Eurycea spp. abundances were influenced by water conductivity. These suggest protection or restoration of riparian habitats and erosion control is an important conservation component for maintaining stream salamanders in the mined landscapes of central Appalachia.

  19. Tooth labeling in cone-beam CT using deep convolutional neural network for forensic identification

    NASA Astrophysics Data System (ADS)

    Miki, Yuma; Muramatsu, Chisako; Hayashi, Tatsuro; Zhou, Xiangrong; Hara, Takeshi; Katsumata, Akitoshi; Fujita, Hiroshi

    2017-03-01

    In large disasters, dental record plays an important role in forensic identification. However, filing dental charts for corpses is not an easy task for general dentists. Moreover, it is laborious and time-consuming work in cases of large scale disasters. We have been investigating a tooth labeling method on dental cone-beam CT images for the purpose of automatic filing of dental charts. In our method, individual tooth in CT images are detected and classified into seven tooth types using deep convolutional neural network. We employed the fully convolutional network using AlexNet architecture for detecting each tooth and applied our previous method using regular AlexNet for classifying the detected teeth into 7 tooth types. From 52 CT volumes obtained by two imaging systems, five images each were randomly selected as test data, and the remaining 42 cases were used as training data. The result showed the tooth detection accuracy of 77.4% with the average false detection of 5.8 per image. The result indicates the potential utility of the proposed method for automatic recording of dental information.

  20. Filtering Gene Ontology semantic similarity for identifying protein complexes in large protein interaction networks.

    PubMed

    Wang, Jian; Xie, Dong; Lin, Hongfei; Yang, Zhihao; Zhang, Yijia

    2012-06-21

    Many biological processes recognize in particular the importance of protein complexes, and various computational approaches have been developed to identify complexes from protein-protein interaction (PPI) networks. However, high false-positive rate of PPIs leads to challenging identification. A protein semantic similarity measure is proposed in this study, based on the ontology structure of Gene Ontology (GO) terms and GO annotations to estimate the reliability of interactions in PPI networks. Interaction pairs with low GO semantic similarity are removed from the network as unreliable interactions. Then, a cluster-expanding algorithm is used to detect complexes with core-attachment structure on filtered network. Our method is applied to three different yeast PPI networks. The effectiveness of our method is examined on two benchmark complex datasets. Experimental results show that our method performed better than other state-of-the-art approaches in most evaluation metrics. The method detects protein complexes from large scale PPI networks by filtering GO semantic similarity. Removing interactions with low GO similarity significantly improves the performance of complex identification. The expanding strategy is also effective to identify attachment proteins of complexes.

  1. Performance of b-jet identification in the ATLAS experiment

    DOE PAGES

    Aad, G; Abbott, B; Abdallah, J; ...

    2016-04-04

    The identification of jets containing b hadrons is important for the physics programme of the ATLAS experiment at the Large Hadron Collider. Several algorithms to identify jets containing b hadrons are described, ranging from those based on the reconstruction of an inclusive secondary vertex or the presence of tracks with large impact parameters to combined tagging algorithms making use of multi-variate discriminants. An independent b-tagging algorithm based on the reconstruction of muons inside jets as well as the b-tagging algorithm used in the online trigger are also presented. The b-jet tagging efficiency, the c-jet tagging efficiency and the mistag ratemore » for light flavour jets in data have been measured with a number of complementary methods. The calibration results are presented as scale factors defined as the ratio of the efficiency (or mistag rate) in data to that in simulation. In the case of b jets, where more than one calibration method exists, the results from the various analyses have been combined taking into account the statistical correlation as well as the correlation of the sources of systematic uncertainty.« less

  2. Multi-innovation auto-constructed least squares identification for 4 DOF ship manoeuvring modelling with full-scale trial data.

    PubMed

    Zhang, Guoqing; Zhang, Xianku; Pang, Hongshuai

    2015-09-01

    This research is concerned with the problem of 4 degrees of freedom (DOF) ship manoeuvring identification modelling with the full-scale trial data. To avoid the multi-innovation matrix inversion in the conventional multi-innovation least squares (MILS) algorithm, a new transformed multi-innovation least squares (TMILS) algorithm is first developed by virtue of the coupling identification concept. And much effort is made to guarantee the uniformly ultimate convergence. Furthermore, the auto-constructed TMILS scheme is derived for the ship manoeuvring motion identification by combination with a statistic index. Comparing with the existing results, the proposed scheme has the significant computational advantage and is able to estimate the model structure. The illustrative examples demonstrate the effectiveness of the proposed algorithm, especially including the identification application with full-scale trial data. Copyright © 2015 ISA. Published by Elsevier Ltd. All rights reserved.

  3. Assessment of management and basic beef quality assurance practices on Idaho dairies.

    PubMed

    Glaze, J B; Chahine, M

    2009-03-01

    In 2004 a mail-in survey was conducted to establish a baseline level of awareness and knowledge related to dairy beef quality assurance (BQA) issues in Idaho. A 30-question survey was mailed to every (n = 736) registered Idaho dairy. Two-hundred seventy-three (37%) dairies participated and were categorized as small (n <201 cows; 53.5%), medium-sized (n = 201 to 1,000 cows; 27.1%) or large (n >1,000 cows; 19.4%). The majority of respondents were dairy owners (83%). Eighty-nine percent of respondents indicated they followed BQA recommendations for animal care. The neck region in cows was used by 68% of respondents for i.m. injections and by 80% for s.c. injections. In calves, the values were 61 and 78%, respectively. Seventy-four percent of respondents indicated they had been trained for injections. Training methods cited included veterinarians (19.8%), dairy owners (16.8%), experience (9.9%), and BQA events or schools (4.5%). The importance of BQA in the dairy industry was rated 2.6 on a 5-point scale (0 = low; 4 = high). Participants rated the effect of dairy animals on the beef industry at 2.5. Plastic ear tags were the preferred method of animal identification, with 100% of large dairies, 97.3% of medium-sized dairies, and 84% of small dairies citing their use. Less than 10% used electronic identification for their animals. Almost half (48%) of large and medium-sized (49%) dairies and 32% of small dairies supported a national animal identification program. A mandatory identification program was supported by 41, 69, and 59% for small, medium-sized, and large dairies, respectively. The percentage of dairies keeping records was similar between small (93%), medium-sized (99%), and large (100%) dairies. Most small dairies (58%) used some form of paper records, whereas most medium-sized (85%) and large (100%) dairies used computers for record keeping. The preferred method to market cull cows by Idaho dairies was the auction market (64%), followed by order buyers (17%), direct to the packer (17%), private treaty sales (16%), and forward contracts (1%). To market calves, dairies used private treaty sales (52%), auction markets (42%), order buyers (14%), and forward contracts (1%). The results of this study will be used by University of Idaho Extension faculty in the design, development, and delivery of dairy BQA program information and materials.

  4. Using Wavelet Analysis To Assist in Identification of Significant Events in Molecular Dynamics Simulations.

    PubMed

    Heidari, Zahra; Roe, Daniel R; Galindo-Murillo, Rodrigo; Ghasemi, Jahan B; Cheatham, Thomas E

    2016-07-25

    Long time scale molecular dynamics (MD) simulations of biological systems are becoming increasingly commonplace due to the availability of both large-scale computational resources and significant advances in the underlying simulation methodologies. Therefore, it is useful to investigate and develop data mining and analysis techniques to quickly and efficiently extract the biologically relevant information from the incredible amount of generated data. Wavelet analysis (WA) is a technique that can quickly reveal significant motions during an MD simulation. Here, the application of WA on well-converged long time scale (tens of μs) simulations of a DNA helix is described. We show how WA combined with a simple clustering method can be used to identify both the physical and temporal locations of events with significant motion in MD trajectories. We also show that WA can not only distinguish and quantify the locations and time scales of significant motions, but by changing the maximum time scale of WA a more complete characterization of these motions can be obtained. This allows motions of different time scales to be identified or ignored as desired.

  5. An intermediate level of abstraction for computational systems chemistry.

    PubMed

    Andersen, Jakob L; Flamm, Christoph; Merkle, Daniel; Stadler, Peter F

    2017-12-28

    Computational techniques are required for narrowing down the vast space of possibilities to plausible prebiotic scenarios, because precise information on the molecular composition, the dominant reaction chemistry and the conditions for that era are scarce. The exploration of large chemical reaction networks is a central aspect in this endeavour. While quantum chemical methods can accurately predict the structures and reactivities of small molecules, they are not efficient enough to cope with large-scale reaction systems. The formalization of chemical reactions as graph grammars provides a generative system, well grounded in category theory, at the right level of abstraction for the analysis of large and complex reaction networks. An extension of the basic formalism into the realm of integer hyperflows allows for the identification of complex reaction patterns, such as autocatalysis, in large reaction networks using optimization techniques.This article is part of the themed issue 'Reconceptualizing the origins of life'. © 2017 The Author(s).

  6. Community concepts of poverty: an application to premium exemptions in Ghana’s National Health Insurance Scheme

    PubMed Central

    2013-01-01

    Background Poverty is multi dimensional. Beyond the quantitative and tangible issues related to inadequate income it also has equally important social, more intangible and difficult if not impossible to quantify dimensions. In 2009, we explored these social and relativist dimension of poverty in five communities in the South of Ghana with differing socio economic characteristics to inform the development and implementation of policies and programs to identify and target the poor for premium exemptions under Ghana’s National Health Insurance Scheme. Methods We employed participatory wealth ranking (PWR) a qualitative tool for the exploration of community concepts, identification and ranking of households into socioeconomic groups. Key informants within the community ranked households into wealth categories after discussing in detail concepts and indicators of poverty. Results Community defined indicators of poverty covered themes related to type of employment, educational attainment of children, food availability, physical appearance, housing conditions, asset ownership, health seeking behavior, social exclusion and marginalization. The poverty indicators discussed shared commonalities but contrasted in the patterns of ranking per community. Conclusion The in-depth nature of the PWR process precludes it from being used for identification of the poor on a large national scale in a program such as the NHIS. However, PWR can provide valuable qualitative input to enrich discussions, development and implementation of policies, programs and tools for large scale interventions and targeting of the poor for social welfare programs such as premium exemption for health care. PMID:23497484

  7. Community concepts of poverty: an application to premium exemptions in Ghana's National Health Insurance Scheme.

    PubMed

    Aryeetey, Genevieve C; Jehu-Appiah, Caroline; Kotoh, Agnes M; Spaan, Ernst; Arhinful, Daniel K; Baltussen, Rob; van der Geest, Sjaak; Agyepong, Irene A

    2013-03-14

    Poverty is multi dimensional. Beyond the quantitative and tangible issues related to inadequate income it also has equally important social, more intangible and difficult if not impossible to quantify dimensions. In 2009, we explored these social and relativist dimension of poverty in five communities in the South of Ghana with differing socio economic characteristics to inform the development and implementation of policies and programs to identify and target the poor for premium exemptions under Ghana's National Health Insurance Scheme. We employed participatory wealth ranking (PWR) a qualitative tool for the exploration of community concepts, identification and ranking of households into socioeconomic groups. Key informants within the community ranked households into wealth categories after discussing in detail concepts and indicators of poverty. Community defined indicators of poverty covered themes related to type of employment, educational attainment of children, food availability, physical appearance, housing conditions, asset ownership, health seeking behavior, social exclusion and marginalization. The poverty indicators discussed shared commonalities but contrasted in the patterns of ranking per community. The in-depth nature of the PWR process precludes it from being used for identification of the poor on a large national scale in a program such as the NHIS. However, PWR can provide valuable qualitative input to enrich discussions, development and implementation of policies, programs and tools for large scale interventions and targeting of the poor for social welfare programs such as premium exemption for health care.

  8. Next-generation ELISA diagnostic assay for Chagas Disease based on the combination of short peptidic epitopes

    PubMed Central

    Volcovich, Romina; Altcheh, Jaime; Bracamonte, Estefanía; Marco, Jorge D.; Nielsen, Morten; Buscaglia, Carlos A.

    2017-01-01

    Chagas Disease, caused by the protozoan Trypanosoma cruzi, is a major health and economic problem in Latin America for which no vaccine or appropriate drugs for large-scale public health interventions are yet available. Accurate diagnosis is essential for the early identification and follow up of vector-borne cases and to prevent transmission of the disease by way of blood transfusions and organ transplantation. Diagnosis is routinely performed using serological methods, some of which require the production of parasite lysates, parasite antigenic fractions or purified recombinant antigens. Although available serological tests give satisfactory results, the production of reliable reagents remains laborious and expensive. Short peptides spanning linear B-cell epitopes have proven ideal serodiagnostic reagents in a wide range of diseases. Recently, we have conducted a large-scale screening of T. cruzi linear B-cell epitopes using high-density peptide chips, leading to the identification of several hundred novel sequence signatures associated to chronic Chagas Disease. Here, we performed a serological assessment of 27 selected epitopes and of their use in a novel multipeptide-based diagnostic method. A combination of 7 of these peptides were finally evaluated in ELISA format against a panel of 199 sera samples (Chagas-positive and negative, including sera from Leishmaniasis-positive subjects). The multipeptide formulation displayed a high diagnostic performance, with a sensitivity of 96.3% and a specificity of 99.15%. Therefore, the use of synthetic peptides as diagnostic tools are an attractive alternative in Chagas’ disease diagnosis. PMID:28991925

  9. Large-Scale Identification and Characterization of Heterodera avenae Putative Effectors Suppressing or Inducing Cell Death in Nicotiana benthamiana

    PubMed Central

    Chen, Changlong; Chen, Yongpan; Jian, Heng; Yang, Dan; Dai, Yiran; Pan, Lingling; Shi, Fengwei; Yang, Shanshan; Liu, Qian

    2018-01-01

    Heterodera avenae is one of the most important plant pathogens and causes vast losses in cereal crops. As a sedentary endoparasitic nematode, H. avenae secretes effectors that modify plant defenses and promote its biotrophic infection of its hosts. However, the number of effectors involved in the interaction between H. avenae and host defenses remains unclear. Here, we report the identification of putative effectors in H. avenae that regulate plant defenses on a large scale. Our results showed that 78 of the 95 putative effectors suppressed programmed cell death (PCD) triggered by BAX and that 7 of the putative effectors themselves caused cell death in Nicotiana benthamiana. Among the cell-death-inducing effectors, three were found to be dependent on their specific domains to trigger cell death and to be expressed in esophageal gland cells by in situ hybridization. Ten candidate effectors that suppressed BAX-triggered PCD also suppressed PCD triggered by the elicitor PsojNIP and at least one R-protein/cognate effector pair, suggesting that they are active in suppressing both pattern-triggered immunity (PTI) and effector-triggered immunity (ETI). Notably, with the exception of isotig16060, these putative effectors could also suppress PCD triggered by cell-death-inducing effectors from H. avenae, indicating that those effectors may cooperate to promote nematode parasitism. Collectively, our results indicate that the majority of the tested effectors of H. avenae may play important roles in suppressing cell death induced by different elicitors in N. benthamiana. PMID:29379510

  10. Multiscale global identification of porous structures

    NASA Astrophysics Data System (ADS)

    Hatłas, Marcin; Beluch, Witold

    2018-01-01

    The paper is devoted to the evolutionary identification of the material constants of porous structures based on measurements conducted on a macro scale. Numerical homogenization with the RVE concept is used to determine the equivalent properties of a macroscopically homogeneous material. Finite element method software is applied to solve the boundary-value problem in both scales. Global optimization methods in form of evolutionary algorithm are employed to solve the identification task. Modal analysis is performed to collect the data necessary for the identification. A numerical example presenting the effectiveness of proposed attitude is attached.

  11. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Nikolic, R J

    This month's issue has the following articles: (1) Dawn of a New Era of Scientific Discovery - Commentary by Edward I. Moses; (2) At the Frontiers of Fundamental Science Research - Collaborators from national laboratories, universities, and international organizations are using the National Ignition Facility to probe key fundamental science questions; (3) Livermore Responds to Crisis in Post-Earthquake Japan - More than 70 Laboratory scientists provided round-the-clock expertise in radionuclide analysis and atmospheric dispersion modeling as part of the nation's support to Japan following the March 2011 earthquake and nuclear accident; (4) A Comprehensive Resource for Modeling, Simulation, and Experimentsmore » - A new Web-based resource called MIDAS is a central repository for material properties, experimental data, and computer models; and (5) Finding Data Needles in Gigabit Haystacks - Livermore computer scientists have developed a novel computer architecture based on 'persistent' memory to ease data-intensive computations.« less

  12. NASA Orbital Debris Baseline Populations

    NASA Technical Reports Server (NTRS)

    Krisko, Paula H.; Vavrin, A. B.

    2013-01-01

    The NASA Orbital Debris Program Office has created high fidelity populations of the debris environment. The populations include objects of 1 cm and larger in Low Earth Orbit through Geosynchronous Transfer Orbit. They were designed for the purpose of assisting debris researchers and sensor developers in planning and testing. This environment is derived directly from the newest ORDEM model populations which include a background derived from LEGEND, as well as specific events such as the Chinese ASAT test, the Iridium 33/Cosmos 2251 accidental collision, the RORSAT sodium-potassium droplet releases, and other miscellaneous events. It is the most realistic ODPO debris population to date. In this paper we present the populations in chart form. We describe derivations of the background population and the specific populations added on. We validate our 1 cm and larger Low Earth Orbit population against SSN, Haystack, and HAX radar measurements.

  13. High-resolution observations of the QSO 3C 345 at 1.3 centimeters

    NASA Technical Reports Server (NTRS)

    Baath, L. B.; Ronnang, B. O.; Pauliny-Toth, I. I. K.; Preuss, E.; Witzel, A.; Matveenko, L. I.; Kogan, L. R.; Kostenko, V. I.; Shaffer, D. B.

    1981-01-01

    High-resolution VLBI observations made at a frequency of 22.235 GHz of the quasar 3C 345 are discussed. Antennas located at the Crimean Astrophysical Observatory, USSR; Onsala, Sweden; Effelsberg, West Germany; and the Haystack Observatory, Massachusetts were employed at 4-min integration times to provide baselines ranging up to 5.5 x 10 to the 8th wavelengths. About 40% of the total flux density of 7.85 Jy, observed in November 1977, and 8.05 Jy, observed in October 1978, is found to originate in an unresolved component of the quasar core in a region less than 0.1 milliarcsec in diameter. The elongated jet-like component of the quasar is observed to contain several peaks of emission extending up to 6 milliarsec from the core which decreased in extent between the two observations.

  14. A Comparative Analysis of Coprologic Diagnostic Methods for Detection of Toxoplama gondii in Cats

    PubMed Central

    Salant, Harold; Spira, Dan T.; Hamburger, Joseph

    2010-01-01

    The relative role of transmission of Toxoplasma gondii infection from cats to humans appears to have recently increased in certain areas. Large-scale screening of oocyst shedding in cats cannot rely on microscopy because oocyst identification lacks sensitivity and specificity, or on bioassays, which require test animals and weeks before examination. We compared a sensitive and species-specific coprologic–polymerase chain reaction (copro-PCR) for detection of T. gondii infected cats with microscopy and a bioassay. In experimentally infected cats followed over time, microscopy was positive occasionally, and positive copro-PCR and bioassay results were obtained continuously from days 2 to 24 post-infection. The copro-PCR is at least as sensitive and specific as the bioassay and is capable of detecting infective oocysts during cat infection. Therefore, this procedure can be used as the new gold standard for determining potential cat infectivity. Its technologic advantages over the bioassay make it superior for large-scale screening of cats. PMID:20439968

  15. Efficient collective influence maximization in cascading processes with first-order transitions

    PubMed Central

    Pei, Sen; Teng, Xian; Shaman, Jeffrey; Morone, Flaviano; Makse, Hernán A.

    2017-01-01

    In many social and biological networks, the collective dynamics of the entire system can be shaped by a small set of influential units through a global cascading process, manifested by an abrupt first-order transition in dynamical behaviors. Despite its importance in applications, efficient identification of multiple influential spreaders in cascading processes still remains a challenging task for large-scale networks. Here we address this issue by exploring the collective influence in general threshold models of cascading process. Our analysis reveals that the importance of spreaders is fixed by the subcritical paths along which cascades propagate: the number of subcritical paths attached to each spreader determines its contribution to global cascades. The concept of subcritical path allows us to introduce a scalable algorithm for massively large-scale networks. Results in both synthetic random graphs and real networks show that the proposed method can achieve larger collective influence given the same number of seeds compared with other scalable heuristic approaches. PMID:28349988

  16. Infrared Multiphoton Dissociation for Quantitative Shotgun Proteomics

    PubMed Central

    Ledvina, Aaron R.; Lee, M. Violet; McAlister, Graeme C.; Westphall, Michael S.; Coon, Joshua J.

    2012-01-01

    We modified a dual-cell linear ion trap mass spectrometer to perform infrared multiphoton dissociation (IRMPD) in the low pressure trap of a dual-cell quadrupole linear ion trap (dual cell QLT) and perform large-scale IRMPD analyses of complex peptide mixtures. Upon optimization of activation parameters (precursor q-value, irradiation time, and photon flux), IRMPD subtly, but significantly outperforms resonant excitation CAD for peptides identified at a 1% false-discovery rate (FDR) from a yeast tryptic digest (95% confidence, p = 0.019). We further demonstrate that IRMPD is compatible with the analysis of isobaric-tagged peptides. Using fixed QLT RF amplitude allows for the consistent retention of reporter ions, but necessitates the use of variable IRMPD irradiation times, dependent upon precursor mass-to-charge (m/z). We show that IRMPD activation parameters can be tuned to allow for effective peptide identification and quantitation simultaneously. We thus conclude that IRMPD performed in a dual-cell ion trap is an effective option for the large-scale analysis of both unmodified and isobaric-tagged peptides. PMID:22480380

  17. Efficient collective influence maximization in cascading processes with first-order transitions

    NASA Astrophysics Data System (ADS)

    Pei, Sen; Teng, Xian; Shaman, Jeffrey; Morone, Flaviano; Makse, Hernán A.

    2017-03-01

    In many social and biological networks, the collective dynamics of the entire system can be shaped by a small set of influential units through a global cascading process, manifested by an abrupt first-order transition in dynamical behaviors. Despite its importance in applications, efficient identification of multiple influential spreaders in cascading processes still remains a challenging task for large-scale networks. Here we address this issue by exploring the collective influence in general threshold models of cascading process. Our analysis reveals that the importance of spreaders is fixed by the subcritical paths along which cascades propagate: the number of subcritical paths attached to each spreader determines its contribution to global cascades. The concept of subcritical path allows us to introduce a scalable algorithm for massively large-scale networks. Results in both synthetic random graphs and real networks show that the proposed method can achieve larger collective influence given the same number of seeds compared with other scalable heuristic approaches.

  18. Large-scale microfluidics providing high-resolution and high-throughput screening of Caenorhabditis elegans poly-glutamine aggregation model

    NASA Astrophysics Data System (ADS)

    Mondal, Sudip; Hegarty, Evan; Martin, Chris; Gökçe, Sertan Kutal; Ghorashian, Navid; Ben-Yakar, Adela

    2016-10-01

    Next generation drug screening could benefit greatly from in vivo studies, using small animal models such as Caenorhabditis elegans for hit identification and lead optimization. Current in vivo assays can operate either at low throughput with high resolution or with low resolution at high throughput. To enable both high-throughput and high-resolution imaging of C. elegans, we developed an automated microfluidic platform. This platform can image 15 z-stacks of ~4,000 C. elegans from 96 different populations using a large-scale chip with a micron resolution in 16 min. Using this platform, we screened ~100,000 animals of the poly-glutamine aggregation model on 25 chips. We tested the efficacy of ~1,000 FDA-approved drugs in improving the aggregation phenotype of the model and identified four confirmed hits. This robust platform now enables high-content screening of various C. elegans disease models at the speed and cost of in vitro cell-based assays.

  19. Ingestion of bacterially expressed double-stranded RNA inhibits gene expression in planarians.

    PubMed

    Newmark, Phillip A; Reddien, Peter W; Cebrià, Francesc; Sánchez Alvarado, Alejandro

    2003-09-30

    Freshwater planarian flatworms are capable of regenerating complete organisms from tiny fragments of their bodies; the basis for this regenerative prowess is an experimentally accessible stem cell population that is present in the adult planarian. The study of these organisms, classic experimental models for investigating metazoan regeneration, has been revitalized by the application of modern molecular biological approaches. The identification of thousands of unique planarian ESTs, coupled with large-scale whole-mount in situ hybridization screens, and the ability to inhibit planarian gene expression through double-stranded RNA-mediated genetic interference, provide a wealth of tools for studying the molecular mechanisms that regulate tissue regeneration and stem cell biology in these organisms. Here we show that, as in Caenorhabditis elegans, ingestion of bacterially expressed double-stranded RNA can inhibit gene expression in planarians. This inhibition persists throughout the process of regeneration, allowing phenotypes with disrupted regenerative patterning to be identified. These results pave the way for large-scale screens for genes involved in regenerative processes.

  20. A coronal hole and its identification as the source of a high velocity solar wind stream

    NASA Technical Reports Server (NTRS)

    Krieger, A. S.; Timothy, A. F.; Roelof, E. C.

    1973-01-01

    X-ray images of the solar corona showed a magnetically open structure in the low corona which extended from N20W20 to the south pole. Analysis of the measured X-ray intensities shows the density scale heights within the structure to be typically a factor of two less than that in the surrounding large scale magnetically closed regions. The structure is identified as a coronal hole. Wind measurements for the appropriate period were traced back to the sun by the method of instantaneous ideal spirals. A striking agreement was found between the Carrington longitude of the solar source of a recurrent high velocity solar wind stream and the position of the hole.

  1. Transitioning Rationally Designed Catalytic Materials to Real 'Working' Catalysts Produced at Commercial Scale: Nanoparticle Materials

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Schaidle, Joshua A.; Habas, Susan E.; Baddour, Frederick G.

    Catalyst design, from idea to commercialization, requires multi-disciplinary scientific and engineering research and development over 10-20 year time periods. Historically, the identification of new or improved catalyst materials has largely been an empirical trial-and-error process. However, advances in computational capabilities (new tools and increased processing power) coupled with new synthetic techniques have started to yield rationally-designed catalysts with controlled nano-structures and tailored properties. This technological advancement represents an opportunity to accelerate the catalyst development timeline and to deliver new materials that outperform existing industrial catalysts or enable new applications, once a number of unique challenges associated with the scale-up ofmore » nano-structured materials are overcome.« less

  2. Seafloor identification in sonar imagery via simulations of Helmholtz equations and discrete optimization

    NASA Astrophysics Data System (ADS)

    Engquist, Björn; Frederick, Christina; Huynh, Quyen; Zhou, Haomin

    2017-06-01

    We present a multiscale approach for identifying features in ocean beds by solving inverse problems in high frequency seafloor acoustics. The setting is based on Sound Navigation And Ranging (SONAR) imaging used in scientific, commercial, and military applications. The forward model incorporates multiscale simulations, by coupling Helmholtz equations and geometrical optics for a wide range of spatial scales in the seafloor geometry. This allows for detailed recovery of seafloor parameters including material type. Simulated backscattered data is generated using numerical microlocal analysis techniques. In order to lower the computational cost of the large-scale simulations in the inversion process, we take advantage of a pre-computed library of representative acoustic responses from various seafloor parameterizations.

  3. Identification of Geostructures of the Continental Crust Particularly as They Relate to Mineral Resource Evaluation. [Alaska

    NASA Technical Reports Server (NTRS)

    Lathram, E. H. (Principal Investigator)

    1974-01-01

    The author has identified the following significant results. A pattern of very old geostructures was recognized, reflecting structures in the crust. This pattern is not peculiar to Alaska, but can be recognized throughout the northern cordillera. A new metallogenic hypothesis for Alaska was developed, based on the relationship of space image linears to known mineral deposits. Using image linear analysis, regional geologic features were also recognized; these features may be used to guide in the location of undiscovered oil and/or gas accumulations in northern Alaska. The effectiveness of ERTS data in enhancing medium and small scale mapping was demonstrated. ERTS data were also used to recognize and monitor the state of large scale vehicular scars on Arctic tundra.

  4. Identification of Gifted Students in Oman: Gender and Grade Differences on the Gifted Rating Scales-School Form

    ERIC Educational Resources Information Center

    Mohamed, Ahmed Hassan Hemdan; Kazem, Ali Mahdi; Pfeiffer, Steven; Alzubaidi, Abdul-Qawi; Elwan, Reda Abu; Ambosaidi, Abdullah; Al-Washahi, Mariam; Al-Kharosi, Tarek

    2017-01-01

    Research suggests that teacher-completed gifted screening scales can reduce undernomination of students with culturally and linguistically diverse backgrounds. The purpose of this study was to examine the use of the Gifted Rating Scales-School Form (GRS-S) in the identification of gifted students in Oman. The participants of the study represented…

  5. Implementation and evaluation of a community-based interprofessional learning activity.

    PubMed

    Luebbers, Ellen L; Dolansky, Mary A; Vehovec, Anton; Petty, Gayle

    2017-01-01

    Implementation of large-scale, meaningful interprofessional learning activities for pre-licensure students has significant barriers and requires novel approaches to ensure success. To accomplish this goal, faculty at Case Western Reserve University, Ohio, USA, used the Ottawa Model of Research Use (OMRU) framework to create, improve, and sustain a community-based interprofessional learning activity for large numbers of medical students (N = 177) and nursing students (N = 154). The model guided the process and included identification of context-specific barriers and facilitators, continual monitoring and improvement using data, and evaluation of student learning outcomes as well as programme outcomes. First year Case Western Reserve University medical students and undergraduate nursing students participated in team-structured prevention screening clinics in the Cleveland Metropolitan Public School District. Identification of barriers and facilitators assisted with overcoming logistic and scheduling issues, large class size, differing ages and skill levels of students and creating sustainability. Continual monitoring led to three distinct phases of improvement and resulted in the creation of an authentic team structure, role clarification, and relevance for students. Evaluation of student learning included both qualitative and quantitative methods, resulting in statistically significant findings and qualitative themes of learner outcomes. The OMRU implementation model provided a useful framework for successful implementation resulting in a sustainable interprofessional learning activity.

  6. Proposed Confidence Scale and ID Score in the Identification of Known-Unknown Compounds Using High Resolution MS Data

    NASA Astrophysics Data System (ADS)

    Rochat, Bertrand

    2017-04-01

    High-resolution (HR) MS instruments recording HR-full scan allow analysts to go further beyond pre-acquisition choices. Untargeted acquisition can reveal unexpected compounds or concentrations and can be performed for preliminary diagnosis attempt. Then, revealed compounds will have to be identified for interpretations. Whereas the need of reference standards is mandatory to confirm identification, the diverse information collected from HRMS allows identifying unknown compounds with relatively high degree of confidence without reference standards injected in the same analytical sequence. However, there is a necessity to evaluate the degree of confidence in putative identifications, possibly before further targeted analyses. This is why a confidence scale and a score in the identification of (non-peptidic) known-unknown, defined as compounds with entries in database, is proposed for (LC-) HRMS data. The scale is based on two representative documents edited by the European Commission (2007/657/EC) and the Metabolomics Standard Initiative (MSI), in an attempt to build a bridge between the communities of metabolomics and screening labs. With this confidence scale, an identification (ID) score is determined as [a number, a letter, and a number] (e.g., 2D3), from the following three criteria: I, a General Identification Category (1, confirmed, 2, putatively identified, 3, annotated compounds/classes, and 4, unknown); II, a Chromatography Class based on the relative retention time (from the narrowest tolerance, A, to no chromatographic references, D); and III, an Identification Point Level (1, very high, 2, high, and 3, normal level) based on the number of identification points collected. Three putative identification examples of known-unknown will be presented.

  7. The consequences of landscape change on ecological resources: An assessment of the United States mid-Atlantic region, 1973-1993

    USGS Publications Warehouse

    Jones, K.B.; Neale, A.C.; Wade, T.G.; Wickham, J.D.; Cross, C.L.; Edmonds, C.M.; Loveland, Thomas R.; Nash, M.S.; Riitters, K.H.; Smith, E.R.

    2001-01-01

    Spatially explicit identification of changes in ecological conditions over large areas is key to targeting and prioritizing areas for environmental protection and restoration by managers at watershed, basin, and regional scales. A critical limitation to this point has been the development of methods to conduct such broad-scale assessments. Field-based methods have proven to be too costly and too inconsistent in their application to make estimates of ecological conditions over large areas. New spatial data derived from satellite imagery and other sources, the development of statistical models relating landscape composition and pattern to ecological endpoints, and geographic information systems (GIS) make it possible to evaluate ecological conditions at multiple scales over broad geographic regions. In this study, we demonstrate the application of spatially distributed models for bird habitat quality and nitrogen yield to streams to assess the consequences of landcover change across the mid-Atlantic region between the 1970s and 1990s. Moreover, we present a way to evaluate spatial concordance between models related to different environmental endpoints. Results of this study should help environmental managers in the mid-Atlantic region target those areas in need of conservation and protection.

  8. Participatory, Multi-Criteria Evaluation Methods as a Means to Increase the Legitimacy and Sustainability of Land Use Planning Processes. The Case of the Chaco Region in Salta, Argentina.

    PubMed

    Seghezzo, Lucas; Venencia, Cristian; Buliubasich, E Catalina; Iribarnegaray, Martín A; Volante, José N

    2017-02-01

    Conflicts over land use and ownership are common in South America and generate frequent confrontations among indigenous peoples, small-scale farmers, and large-scale agricultural producers. We argue in this paper that an accurate identification of these conflicts, together with a participatory evaluation of their importance, will increase the social legitimacy of land use planning processes, rendering decision-making more sustainable in the long term. We describe here a participatory, multi-criteria conflict assessment model developed to identify, locate, and categorize land tenure and use conflicts. The model was applied to the case of the "Chaco" region of the province of Salta, in northwestern Argentina. Basic geographic, cadastral, and social information needed to apply the model was made spatially explicit on a Geographic Information System. Results illustrate the contrasting perceptions of different stakeholders (government officials, social and environmental non-governmental organizations, large-scale agricultural producers, and scholars) on the intensity of land use conflicts in the study area. These results can help better understand and address land tenure conflicts in areas with different cultures and conflicting social and enviornmental interests.

  9. Participatory, Multi-Criteria Evaluation Methods as a Means to Increase the Legitimacy and Sustainability of Land Use Planning Processes. The Case of the Chaco Region in Salta, Argentina

    NASA Astrophysics Data System (ADS)

    Seghezzo, Lucas; Venencia, Cristian; Buliubasich, E. Catalina; Iribarnegaray, Martín A.; Volante, José N.

    2017-02-01

    Conflicts over land use and ownership are common in South America and generate frequent confrontations among indigenous peoples, small-scale farmers, and large-scale agricultural producers. We argue in this paper that an accurate identification of these conflicts, together with a participatory evaluation of their importance, will increase the social legitimacy of land use planning processes, rendering decision-making more sustainable in the long term. We describe here a participatory, multi-criteria conflict assessment model developed to identify, locate, and categorize land tenure and use conflicts. The model was applied to the case of the "Chaco" region of the province of Salta, in northwestern Argentina. Basic geographic, cadastral, and social information needed to apply the model was made spatially explicit on a Geographic Information System. Results illustrate the contrasting perceptions of different stakeholders (government officials, social and environmental non-governmental organizations, large-scale agricultural producers, and scholars) on the intensity of land use conflicts in the study area. These results can help better understand and address land tenure conflicts in areas with different cultures and conflicting social and enviornmental interests.

  10. The Problem of Late ART initiation in Sub-Saharan Africa: A Transient Aspect of Scale-up or a Long-term Phenomenon?

    PubMed Central

    Lahuerta, Maria; Ue, Frances; Hoffman, Susie; Elul, Batya; Kulkarni, Sarah Gorrell; Wu, Yingfeng; Nuwagaba-Biribonwoha, Harriet; Remien, Robert H.; Sadr, Wafaa El; Nash, Denis

    2013-01-01

    Efforts to scale-up HIV care and treatment have been successful at initiating large numbers of patients onto antiretroviral therapy (ART), although persistent challenges remain to optimizing scale-up effectiveness in both resource-rich and resource-limited settings. Among the most important are very high rates of ART initiation in the advanced stages of HIV disease, which in turn drive morbidity, mortality, and onward transmission of HIV. With a focus on sub-Saharan Africa, this review article presents a conceptual framework for a broader discussion of the persistent problem of late ART initiation, including a need for more focus on the upstream precursors (late HIV diagnosis and late enrollment into HIV care) and their determinants. Without additional research and identification of multilevel interventions that successfully promote earlier initiation of ART, the problem of late ART initiation will persist, significantly undermining the long-term impact of HIV care scale-up on reducing mortality and controlling the HIV epidemic. PMID:23377739

  11. Assessment of automatic ligand building in ARP/wARP.

    PubMed

    Evrard, Guillaume X; Langer, Gerrit G; Perrakis, Anastassis; Lamzin, Victor S

    2007-01-01

    The efficiency of the ligand-building module of ARP/wARP version 6.1 has been assessed through extensive tests on a large variety of protein-ligand complexes from the PDB, as available from the Uppsala Electron Density Server. Ligand building in ARP/wARP involves two main steps: automatic identification of the location of the ligand and the actual construction of its atomic model. The first step is most successful for large ligands. The second step, ligand construction, is more powerful with X-ray data at high resolution and ligands of small to medium size. Both steps are successful for ligands with low to moderate atomic displacement parameters. The results highlight the strengths and weaknesses of both the method of ligand building and the large-scale validation procedure and help to identify means of further improvement.

  12. Volumetric three-component velocimetry measurements of the turbulent flow around a Rushton turbine

    NASA Astrophysics Data System (ADS)

    Sharp, Kendra V.; Hill, David; Troolin, Daniel; Walters, Geoffrey; Lai, Wing

    2010-01-01

    Volumetric three-component velocimetry measurements have been taken of the flow field near a Rushton turbine in a stirred tank reactor. This particular flow field is highly unsteady and three-dimensional, and is characterized by a strong radial jet, large tank-scale ring vortices, and small-scale blade tip vortices. The experimental technique uses a single camera head with three apertures to obtain approximately 15,000 three-dimensional vectors in a cubic volume. These velocity data offer the most comprehensive view to date of this flow field, especially since they are acquired at three Reynolds numbers (15,000, 107,000, and 137,000). Mean velocity fields and turbulent kinetic energy quantities are calculated. The volumetric nature of the data enables tip vortex identification, vortex trajectory analysis, and calculation of vortex strength. Three identification methods for the vortices are compared based on: the calculation of circumferential vorticity; the calculation of local pressure minima via an eigenvalue approach; and the calculation of swirling strength again via an eigenvalue approach. The use of two-dimensional data and three-dimensional data is compared for vortex identification; a `swirl strength' criterion is less sensitive to completeness of the velocity gradient tensor and overall provides clearer identification of the tip vortices. The principal components of the strain rate tensor are also calculated for one Reynolds number case as these measures of stretching and compression have recently been associated with tip vortex characterization. Vortex trajectories and strength compare favorably with those in the literature. No clear dependence of trajectory on Reynolds number is deduced. The visualization of tip vortices up to 140° past blade passage in the highest Reynolds number case is notable and has not previously been shown.

  13. Finding a Needle in a Haystack: Distinguishing Mexican Maize Landraces Using a Small Number of SNPs

    PubMed Central

    Caldu-Primo, Jose L.; Mastretta-Yanes, Alicia; Wegier, Ana; Piñero, Daniel

    2017-01-01

    In Mexico's territory, the center of origin and domestication of maize (Zea mays), there is a large phenotypic diversity of this crop. This diversity has been classified into “landraces.” Previous studies have reported that genomic variation in Mexican maize is better explained by environmental factors, particularly those related with altitude, than by landrace. Still, landraces are extensively used by agronomists, who recognize them as stable and discriminatory categories for the classification of samples. In order to investigate the genomic foundation of maize landraces, we analyzed genomic data (35,909 SNPs from Illumina MaizeSNP50 BeadChip) obtained from 50 samples representing five maize landraces (Comiteco, Conejo, Tehua, Zapalote Grande, and Zapalote Chico), and searched for markers suitable for landrace assignment. Landrace clusters could not be identified taking all the genomic information, but they become manifest taking only a subset of SNPs with high FST among landraces. Discriminant analysis of principal components was conducted to classify samples using SNP data. Two classification analyses were done, first classifying samples by landrace and then by altitude category. Through this classification method, we identified 20 landrace-informative SNPs and 14 altitude-informative SNPs, with only 6 SNPs in common for both analyses. These results show that Mexican maize phenotypic diversity can be classified in landraces using a small number of genomic markers, given the fact that landrace genomic diversity is influenced by environmental factors as well as artificial selection due to bio-cultural practices. PMID:28458682

  14. Describing Ecosystem Complexity through Integrated Catchment Modeling

    NASA Astrophysics Data System (ADS)

    Shope, C. L.; Tenhunen, J. D.; Peiffer, S.

    2011-12-01

    Land use and climate change have been implicated in reduced ecosystem services (ie: high quality water yield, biodiversity, and agricultural yield. The prediction of ecosystem services expected under future land use decisions and changing climate conditions has become increasingly important. Complex policy and management decisions require the integration of physical, economic, and social data over several scales to assess effects on water resources and ecology. Field-based meteorology, hydrology, soil physics, plant production, solute and sediment transport, economic, and social behavior data were measured in a South Korean catchment. A variety of models are being used to simulate plot and field scale experiments within the catchment. Results from each of the local-scale models provide identification of sensitive, local-scale parameters which are then used as inputs into a large-scale watershed model. We used the spatially distributed SWAT model to synthesize the experimental field data throughout the catchment. The approach of our study was that the range in local-scale model parameter results can be used to define the sensitivity and uncertainty in the large-scale watershed model. Further, this example shows how research can be structured for scientific results describing complex ecosystems and landscapes where cross-disciplinary linkages benefit the end result. The field-based and modeling framework described is being used to develop scenarios to examine spatial and temporal changes in land use practices and climatic effects on water quantity, water quality, and sediment transport. Development of accurate modeling scenarios requires understanding the social relationship between individual and policy driven land management practices and the value of sustainable resources to all shareholders.

  15. Deep learning with non-medical training used for chest pathology identification

    NASA Astrophysics Data System (ADS)

    Bar, Yaniv; Diamant, Idit; Wolf, Lior; Greenspan, Hayit

    2015-03-01

    In this work, we examine the strength of deep learning approaches for pathology detection in chest radiograph data. Convolutional neural networks (CNN) deep architecture classification approaches have gained popularity due to their ability to learn mid and high level image representations. We explore the ability of a CNN to identify different types of pathologies in chest x-ray images. Moreover, since very large training sets are generally not available in the medical domain, we explore the feasibility of using a deep learning approach based on non-medical learning. We tested our algorithm on a dataset of 93 images. We use a CNN that was trained with ImageNet, a well-known large scale nonmedical image database. The best performance was achieved using a combination of features extracted from the CNN and a set of low-level features. We obtained an area under curve (AUC) of 0.93 for Right Pleural Effusion detection, 0.89 for Enlarged heart detection and 0.79 for classification between healthy and abnormal chest x-ray, where all pathologies are combined into one large class. This is a first-of-its-kind experiment that shows that deep learning with large scale non-medical image databases may be sufficient for general medical image recognition tasks.

  16. Initial velocity V-shapes of young asteroid families

    NASA Astrophysics Data System (ADS)

    Bolin, Bryce T.; Walsh, Kevin J.; Morbidelli, Alessandro; Delbó, Marco

    2018-01-01

    Ejection velocity fields of asteroid families are largely unconstrained due to the fact that members disperse relatively quickly on Myr time-scales by secular resonances and the Yarkovsky effect. The spreading of fragments in a by the Yarkovsky effect is indistinguishable from the spreading caused by the initial ejection of fragments. By examining families <20 Myr old, we can use the V-shape identification technique to separate family shapes that are due to the initial ejection velocity field and those that are due to the Yarkovsky effect. Asteroid families that are <20 Myr old provide an opportunity to study the velocity field of family fragments before they become too dispersed. Only the Karin family's initial velocity field has been determined and scales inversely with diameter, D-1. We have applied the V-shape identification technique to constrain young families' initial ejection velocity fields by measuring the curvature of their fragments' V-shape correlation in semimajor axis, a, versus D-1 space. Curvature from a straight line implies a deviation from a scaling of D-1. We measure the V-shape curvature of 11 young asteroid families including the 1993 FY12, Aeolia, Brangane, Brasilia, Clarissa, Iannini, Karin, Konig, Koronis(2), Theobalda and Veritas asteroid families. We find that the majority of asteroid families have initial ejection velocity fields consistent with ∼D-1 supporting laboratory impact experiments and computer simulations of disrupting asteroid parent bodies.

  17. Structural similitude and design of scaled down laminated models

    NASA Technical Reports Server (NTRS)

    Simitses, G. J.; Rezaeepazhand, J.

    1993-01-01

    The excellent mechanical properties of laminated composite structures make them prime candidates for wide variety of applications in aerospace, mechanical and other branches of engineering. The enormous design flexibility of advanced composites is obtained at the cost of large number of design parameters. Due to complexity of the systems and lack of complete design based informations, designers tend to be conservative in their design. Furthermore, any new design is extensively evaluated experimentally until it achieves the necessary reliability, performance and safety. However, the experimental evaluation of composite structures are costly and time consuming. Consequently, it is extremely useful if a full-scale structure can be replaced by a similar scaled-down model which is much easier to work with. Furthermore, a dramatic reduction in cost and time can be achieved, if available experimental data of a specific structure can be used to predict the behavior of a group of similar systems. This study investigates problems associated with the design of scaled models. Such study is important since it provides the necessary scaling laws, and the factors which affect the accuracy of the scale models. Similitude theory is employed to develop the necessary similarity conditions (scaling laws). Scaling laws provide relationship between a full-scale structure and its scale model, and can be used to extrapolate the experimental data of a small, inexpensive, and testable model into design information for a large prototype. Due to large number of design parameters, the identification of the principal scaling laws by conventional method (dimensional analysis) is tedious. Similitude theory based on governing equations of the structural system is more direct and simpler in execution. The difficulty of making completely similar scale models often leads to accept certain type of distortion from exact duplication of the prototype (partial similarity). Both complete and partial similarity are discussed. The procedure consists of systematically observing the effect of each parameter and corresponding scaling laws. Then acceptable intervals and limitations for these parameters and scaling laws are discussed. In each case, a set of valid scaling factors and corresponding response scaling laws that accurately predict the response of prototypes from experimental models is introduced. The examples used include rectangular laminated plates under destabilizing loads, applied individually, vibrational characteristics of same plates, as well as cylindrical bending of beam-plates.

  18. Stepwise identification of HLA-A*0201-restricted CD8+ T-cell epitope peptides from herpes simplex virus type 1 genome boosted by a StepRank scheme.

    PubMed

    Bi, Jianjun; Song, Rengang; Yang, Huilan; Li, Bingling; Fan, Jianyong; Liu, Zhongrong; Long, Chaoqin

    2011-01-01

    Identification of immunodominant epitopes is the first step in the rational design of peptide vaccines aimed at T-cell immunity. To date, however, it is yet a great challenge for accurately predicting the potent epitope peptides from a pool of large-scale candidates with an efficient manner. In this study, a method that we named StepRank has been developed for the reliable and rapid prediction of binding capabilities/affinities between proteins and genome-wide peptides. In this procedure, instead of single strategy used in most traditional epitope identification algorithms, four steps with different purposes and thus different computational demands are employed in turn to screen the large-scale peptide candidates that are normally generated from, for example, pathogenic genome. The steps 1 and 2 aim at qualitative exclusion of typical nonbinders by using empirical rule and linear statistical approach, while the steps 3 and 4 focus on quantitative examination and prediction of the interaction energy profile and binding affinity of peptide to target protein via quantitative structure-activity relationship (QSAR) and structure-based free energy analysis. We exemplify this method through its application to binding predictions of the peptide segments derived from the 76 known open-reading frames (ORFs) of herpes simplex virus type 1 (HSV-1) genome with or without affinity to human major histocompatibility complex class I (MHC I) molecule HLA-A*0201, and find that the predictive results are well compatible with the classical anchor residue theory and perfectly match for the extended motif pattern of MHC I-binding peptides. The putative epitopes are further confirmed by comparisons with 11 experimentally measured HLA-A*0201-restrcited peptides from the HSV-1 glycoproteins D and K. We expect that this well-designed scheme can be applied in the computational screening of other viral genomes as well.

  19. System identification through nonstationary data using Time-Frequency Blind Source Separation

    NASA Astrophysics Data System (ADS)

    Guo, Yanlin; Kareem, Ahsan

    2016-06-01

    Classical output-only system identification (SI) methods are based on the assumption of stationarity of the system response. However, measured response of buildings and bridges is usually non-stationary due to strong winds (e.g. typhoon, and thunder storm etc.), earthquakes and time-varying vehicle motions. Accordingly, the response data may have time-varying frequency contents and/or overlapping of modal frequencies due to non-stationary colored excitation. This renders traditional methods problematic for modal separation and identification. To address these challenges, a new SI technique based on Time-Frequency Blind Source Separation (TFBSS) is proposed. By selectively utilizing "effective" information in local regions of the time-frequency plane, where only one mode contributes to energy, the proposed technique can successfully identify mode shapes and recover modal responses from the non-stationary response where the traditional SI methods often encounter difficulties. This technique can also handle response with closely spaced modes which is a well-known challenge for the identification of large-scale structures. Based on the separated modal responses, frequency and damping can be easily identified using SI methods based on a single degree of freedom (SDOF) system. In addition to the exclusive advantage of handling non-stationary data and closely spaced modes, the proposed technique also benefits from the absence of the end effects and low sensitivity to noise in modal separation. The efficacy of the proposed technique is demonstrated using several simulation based studies, and compared to the popular Second-Order Blind Identification (SOBI) scheme. It is also noted that even some non-stationary response data can be analyzed by the stationary method SOBI. This paper also delineates non-stationary cases where SOBI and the proposed scheme perform comparably and highlights cases where the proposed approach is more advantageous. Finally, the performance of the proposed method is evaluated using a full-scale non-stationary response of a tall building during an earthquake and found it to perform satisfactorily.

  20. Identification of scintillation signatures on GPS signals originating from plasma structures detected with EISCAT incoherent scatter radar along the same line of sight

    NASA Astrophysics Data System (ADS)

    Forte, Biagio; Coleman, Chris; Skone, Susan; Häggström, Ingemar; Mitchell, Cathryn; Da Dalt, Federico; Panicciari, Tommaso; Kinrade, Joe; Bust, Gary

    2017-01-01

    Ionospheric scintillation originates from the scattering of electromagnetic waves through spatial gradients in the plasma density distribution, drifting across a given propagation direction. Ionospheric scintillation represents a disruptive manifestation of adverse space weather conditions through degradation of the reliability and continuity of satellite telecommunication and navigation systems and services (e.g., European Geostationary Navigation Overlay Service, EGNOS). The purpose of the experiment presented here was to determine the contribution of auroral ionization structures to GPS scintillation. European Incoherent Scatter (EISCAT) measurements were obtained along the same line of sight of a given GPS satellite observed from Tromso and followed by means of the EISCAT UHF radar to causally identify plasma structures that give rise to scintillation on the co-aligned GPS radio link. Large-scale structures associated with the poleward edge of the ionospheric trough, with auroral arcs in the nightside auroral oval and with particle precipitation at the onset of a substorm were indeed identified as responsible for enhanced phase scintillation at L band. For the first time it was observed that the observed large-scale structures did not cascade into smaller-scale structures, leading to enhanced phase scintillation without amplitude scintillation. More measurements and theory are necessary to understand the mechanism responsible for the inhibition of large-scale to small-scale energy cascade and to reproduce the observations. This aspect is fundamental to model the scattering of radio waves propagating through these ionization structures. New insights from this experiment allow a better characterization of the impact that space weather can have on satellite telecommunications and navigation services.

  1. Identification of scintillation signatures on GPS signals originating from plasma structures detected with EISCAT incoherent scatter radar along the same line of sight.

    PubMed

    Forte, Biagio; Coleman, Chris; Skone, Susan; Häggström, Ingemar; Mitchell, Cathryn; Da Dalt, Federico; Panicciari, Tommaso; Kinrade, Joe; Bust, Gary

    2017-01-01

    Ionospheric scintillation originates from the scattering of electromagnetic waves through spatial gradients in the plasma density distribution, drifting across a given propagation direction. Ionospheric scintillation represents a disruptive manifestation of adverse space weather conditions through degradation of the reliability and continuity of satellite telecommunication and navigation systems and services (e.g., European Geostationary Navigation Overlay Service, EGNOS). The purpose of the experiment presented here was to determine the contribution of auroral ionization structures to GPS scintillation. European Incoherent Scatter (EISCAT) measurements were obtained along the same line of sight of a given GPS satellite observed from Tromso and followed by means of the EISCAT UHF radar to causally identify plasma structures that give rise to scintillation on the co-aligned GPS radio link. Large-scale structures associated with the poleward edge of the ionospheric trough, with auroral arcs in the nightside auroral oval and with particle precipitation at the onset of a substorm were indeed identified as responsible for enhanced phase scintillation at L band. For the first time it was observed that the observed large-scale structures did not cascade into smaller-scale structures, leading to enhanced phase scintillation without amplitude scintillation. More measurements and theory are necessary to understand the mechanism responsible for the inhibition of large-scale to small-scale energy cascade and to reproduce the observations. This aspect is fundamental to model the scattering of radio waves propagating through these ionization structures. New insights from this experiment allow a better characterization of the impact that space weather can have on satellite telecommunications and navigation services.

  2. Study for identification of beneficial uses of Space (BUS). Volume 2: Technical report. Book 1: Development and business analysis of space processed isoenzymes

    NASA Technical Reports Server (NTRS)

    1975-01-01

    A separation method to provide reasonable yields of high specificity isoenzymes for the purpose of large scale, early clinical diagnosis of diseases and organic damage such as, myocardial infarction, hepatoma, muscular dystrophy, and infectous disorders is presented. Preliminary development plans are summarized. An analysis of required research and development and production resources is included. The costs of such resources and the potential profitability of a commercial space processing opportunity for electrophoretic separation of high specificity isoenzymes are reviewed.

  3. Smart sensors II; Proceedings of the Seminar, San Diego, CA, July 31, August 1, 1980

    NASA Astrophysics Data System (ADS)

    Barbe, D. F.

    1980-01-01

    Topics discussed include technology for smart sensors, smart sensors for tracking and surveillance, and techniques and algorithms for smart sensors. Papers are presented on the application of very large scale integrated circuits to smart sensors, imaging charge-coupled devices for deep-space surveillance, ultra-precise star tracking using charge coupled devices, and automatic target identification of blurred images with super-resolution features. Attention is also given to smart sensors for terminal homing, algorithms for estimating image position, and the computational efficiency of multiple image registration algorithms.

  4. Identification of Curie temperature distributions in magnetic particulate systems

    NASA Astrophysics Data System (ADS)

    Waters, J.; Berger, A.; Kramer, D.; Fangohr, H.; Hovorka, O.

    2017-09-01

    This paper develops a methodology for extracting the Curie temperature distribution from magnetisation versus temperature measurements which are realizable by standard laboratory magnetometry. The method is integral in nature, robust against various sources of measurement noise, and can be adopted to a wide range of granular magnetic materials and magnetic particle systems. The validity and practicality of the method is demonstrated using large-scale Monte-Carlo simulations of an Ising-like model as a proof of concept, and general conclusions are drawn about its applicability to different classes of systems and experimental conditions.

  5. System for analysis of LANDSAT agricultural data: Automatic computer-assisted proportion estimation of local areas

    NASA Technical Reports Server (NTRS)

    Nalepka, R. F. (Principal Investigator); Kauth, R. J.; Thomas, G. S.

    1976-01-01

    The author has identified the following significant results. A conceptual man machine system framework was created for a large scale agricultural remote sensing system. The system is based on and can grow out of the local recognition mode of LACIE, through a gradual transition wherein computer support functions supplement and replace AI functions. Local proportion estimation functions are broken into two broad classes: (1) organization of the data within the sample segment; and (2) identification of the fields or groups of fields in the sample segment.

  6. Use of Machine Learning Techniques for Identification of Robust Teleconnections to East African Rainfall Variability

    NASA Technical Reports Server (NTRS)

    Roberts, J. Brent; Robertson, F. R.; Funk, C.

    2014-01-01

    Hidden Markov models can be used to investigate structure of subseasonal variability. East African short rain variability has connections to large-scale tropical variability. MJO - Intraseasonal variations connected with appearance of "wet" and "dry" states. ENSO/IOZM SST and circulation anomalies are apparent during years of anomalous residence time in the subseasonal "wet" state. Similar results found in previous studies, but we can interpret this with respect to variations of subseasonal wet and dry modes. Reveal underlying connections between MJO/IOZM/ENSO with respect to East African rainfall.

  7. Absolute pitch among students at the Shanghai Conservatory of Music: a large-scale direct-test study.

    PubMed

    Deutsch, Diana; Li, Xiaonuo; Shen, Jing

    2013-11-01

    This paper reports a large-scale direct-test study of absolute pitch (AP) in students at the Shanghai Conservatory of Music. Overall note-naming scores were very high, with high scores correlating positively with early onset of musical training. Students who had begun training at age ≤5 yr scored 83% correct not allowing for semitone errors and 90% correct allowing for semitone errors. Performance levels were higher for white key pitches than for black key pitches. This effect was greater for orchestral performers than for pianists, indicating that it cannot be attributed to early training on the piano. Rather, accuracy in identifying notes of different names (C, C#, D, etc.) correlated with their frequency of occurrence in a large sample of music taken from the Western tonal repertoire. There was also an effect of pitch range, so that performance on tones in the two-octave range beginning on Middle C was higher than on tones in the octave below Middle C. In addition, semitone errors tended to be on the sharp side. The evidence also ran counter to the hypothesis, previously advanced by others, that the note A plays a special role in pitch identification judgments.

  8. Association between major depressive disorder and odor identification impairment.

    PubMed

    Khil, Laura; Rahe, Corinna; Wellmann, Jürgen; Baune, Bernhard T; Wersching, Heike; Berger, Klaus

    2016-10-01

    There is evidence of olfactory deficits in patients with major depressive disorder (MDD) but causes and mechanisms are largely unknown. We compared 728 patients with current MDD and 555 non-depressed controls regarding odor identification impairment taking into account the severity of acute symptoms and of the disease course. We assessed current symptom severity with the Hamilton Depression Rating Scale, and disease course severity based on admission diagnosis (ICD-10, F32/F33) and self-reported hospitalization frequency, defined as infrequent (<2) and frequent (≥2) depression-related hospitalizations under constant disease duration. A score of <10 on the Sniffin' Sticks-Screen-12 test determined the presence of odor identification impairment. Compared to non-depressed controls patients with frequent (rapidly recurring) hospitalizations had an elevated chance of odor identification impairment, even after adjustment for smell-influencing factors, such as age and smoking, (OR=1.7; 95% CI 1.0-2.9). Patients with recurrent MDD (F33) also had an elevated odds of odor identification impairment compared to those with a first-time episode (F32, OR=1.5; 95% CI 1.0-2.4). In patients with a first-time episode the chance of odor identification impairment increased by 7% with each point increase in the Hamilton Score. Cross-sectional study. Variation in the use of psychotropic medication is a potential bias. Odor identification impairment was evident in MDD patients with first-time high symptom severity and in patients with a severe disease course. Whether odor identification impairment is a marker or mediator of structural and functional brain changes associated with acute or active MDD requires further investigations in longitudinal studies. Copyright © 2016 Elsevier B.V. All rights reserved.

  9. Relationship of dysfunctional sport fandom with dislike for rivals in a sample of college students.

    PubMed

    Smith, Jana; Wann, Daniel L

    2006-06-01

    The relationships among sport-fandom dysfunctionality (tendencies toward complaining and confrontation as assessed via the Dysfunctional Sport Fandom Scale) and items assessing team identification (assessed via the Sport Spectator Identification Scale) were examined with 87 college students (24 men, 63 women, M age=20.2 yr.). Although positive associations of dysfunction and identification were found, contrary to expectations, the relationship between dysfunction and dislike for rivals was not particularly strong.

  10. Revisiting Cholera-Climate Teleconnections in the Native Homeland: ENSO and other Extremes through the Regional Hydroclimatic Drivers

    NASA Astrophysics Data System (ADS)

    Akanda, A. S.; Jutla, A.; Huq, A.; Colwell, R. R.

    2014-12-01

    Cholera is a global disease, with significantly large outbreaks occurring since the 1990s, notably in Sub-Saharan Africa and South Asia and recently in Haiti, in the Caribbean. Critical knowledge gaps remain in the understanding of the annual recurrence in endemic areas and the nature of epidemic outbreaks, especially those that follow extreme hydroclimatic events. Teleconnections with large-scale climate phenomena affecting regional scale hydroclimatic drivers of cholera dynamics remain largely unexplained. For centuries, the Bengal delta region has been strongly influenced by the asymmetric availability of water in the rivers Ganges and the Brahmaputra. As these two major rivers are known to have strong contrasting affects on local cholera dynamics in the region, we argue that the role of El Nino-Southern Oscillation (ENSO), Indian Ocean Dipole (IOD), or other phenomena needs to be interpreted in the context of the seasonal role of individual rivers and subsequent impact on local environmental processes, not as a teleconnection having a remote and unified effect. We present a modified hypothesis that the influences of large-scale climate phenomena such as ENSO and IOD on Bengal cholera can be explicitly identified and incorporated through regional scale hydroclimatic drivers. Here, we provide an analytical review of the literature addressing cholera and climate linkages and present hypotheses, based on recent evidence, and quantification on the role of regional scale hydroclimatic drivers of cholera. We argue that the seasonal changes in precipitation and temperature, and resulting river discharge in the GBM basin region during ENSO and IOD events have a dominant combined effect on the endemic persistence and the epidemic vulnerability to cholera outbreaks in spring and fall seasons, respectively, that is stronger than the effect of localized hydrological and socio-economic sensitivities in Bangladesh. In addition, systematic identification of underlying seasonal hydroclimatic drivers will allow us to harness the inherent system memory of these processes to develop early warning systems and strengthen prevention measures.

  11. Sorting Through the Safety Data Haystack: Using Machine Learning to Identify Individual Case Safety Reports in Social-Digital Media.

    PubMed

    Comfort, Shaun; Perera, Sujan; Hudson, Zoe; Dorrell, Darren; Meireis, Shawman; Nagarajan, Meenakshi; Ramakrishnan, Cartic; Fine, Jennifer

    2018-06-01

    There is increasing interest in social digital media (SDM) as a data source for pharmacovigilance activities; however, SDM is considered a low information content data source for safety data. Given that pharmacovigilance itself operates in a high-noise, lower-validity environment without objective 'gold standards' beyond process definitions, the introduction of large volumes of SDM into the pharmacovigilance workflow has the potential to exacerbate issues with limited manual resources to perform adverse event identification and processing. Recent advances in medical informatics have resulted in methods for developing programs which can assist human experts in the detection of valid individual case safety reports (ICSRs) within SDM. In this study, we developed rule-based and machine learning (ML) models for classifying ICSRs from SDM and compared their performance with that of human pharmacovigilance experts. We used a random sampling from a collection of 311,189 SDM posts that mentioned Roche products and brands in combination with common medical and scientific terms sourced from Twitter, Tumblr, Facebook, and a spectrum of news media blogs to develop and evaluate three iterations of an automated ICSR classifier. The ICSR classifier models consisted of sub-components to annotate the relevant ICSR elements and a component to make the final decision on the validity of the ICSR. Agreement with human pharmacovigilance experts was chosen as the preferred performance metric and was evaluated by calculating the Gwet AC1 statistic (gKappa). The best performing model was tested against the Roche global pharmacovigilance expert using a blind dataset and put through a time test of the full 311,189-post dataset. During this effort, the initial strict rule-based approach to ICSR classification resulted in a model with an accuracy of 65% and a gKappa of 46%. Adding an ML-based adverse event annotator improved the accuracy to 74% and gKappa to 60%. This was further improved by the addition of an additional ML ICSR detector. On a blind test set of 2500 posts, the final model demonstrated a gKappa of 78% and an accuracy of 83%. In the time test, it took the final model 48 h to complete a task that would have taken an estimated 44,000 h for human experts to perform. The results of this study indicate that an effective and scalable solution to the challenge of ICSR detection in SDM includes a workflow using an automated ML classifier to identify likely ICSRs for further human SME review.

  12. Olfactory identification deficit and its relationship with hedonic traits in patients with first-episode schizophrenia and individuals with schizotypy.

    PubMed

    Zou, Lai-Quan; Zhou, Han-Yu; Lui, Simon S Y; Wang, Yi; Wang, Ya; Gan, Jun; Zhu, Xiong-Zhao; Cheung, Eric F C; Chan, Raymond C K

    2018-04-20

    Olfactory identification impairments have been consistently found in schizophrenia patients. However, few previous studies have investigated this in first-episode patients. There are also inconsistent findings regarding olfactory identification ability in psychometrically-defined schizotypy individuals. In this study, we directly compared the olfactory identification ability of first-episode schizophrenia patients with schizotypy individuals. The relationship between olfactory identification impairments and hedonic traits was also examined. Thirty-five first-episode schizophrenia patients, 40 schizotypy individuals as defined by the Chapman's Anhedonia Scales and 40 demographically matched controls were recruited. The University of Pennsylvania Smell Identification Test was administered. Hedonic capacity was assessed using the Temporal Experience of Pleasure Scale (TEPS). The results showed that both the schizophrenia and schizotypy groups showed poorer olfactory identification ability than controls, and the impairment was significantly correlated with reduced pleasure experiences. Our findings support olfactory identification impairment as a trait marker for schizophrenia. Copyright © 2018 Elsevier Inc. All rights reserved.

  13. Dropout Proneness in Appalachia. Research Series 3.

    ERIC Educational Resources Information Center

    Mink, Oscar G.; Barker, Laurence W.

    Two aids used in the identification of potential dropouts are examined. The Mink Scale (a teacher-rated scale) is based on classification of social, psychological, and educational forces related to dropout proneness: (1) academic ability and performance, (2) negative identification with education, (3) family and socioeconomic status, and (4)…

  14. XLID-Causing Mutations and Associated Genes Challenged in Light of Data From Large-Scale Human Exome Sequencing

    PubMed Central

    Piton, Amélie; Redin, Claire; Mandel, Jean-Louis

    2013-01-01

    Because of the unbalanced sex ratio (1.3–1.4 to 1) observed in intellectual disability (ID) and the identification of large ID-affected families showing X-linked segregation, much attention has been focused on the genetics of X-linked ID (XLID). Mutations causing monogenic XLID have now been reported in over 100 genes, most of which are commonly included in XLID diagnostic gene panels. Nonetheless, the boundary between true mutations and rare non-disease-causing variants often remains elusive. The sequencing of a large number of control X chromosomes, required for avoiding false-positive results, was not systematically possible in the past. Such information is now available thanks to large-scale sequencing projects such as the National Heart, Lung, and Blood (NHLBI) Exome Sequencing Project, which provides variation information on 10,563 X chromosomes from the general population. We used this NHLBI cohort to systematically reassess the implication of 106 genes proposed to be involved in monogenic forms of XLID. We particularly question the implication in XLID of ten of them (AGTR2, MAGT1, ZNF674, SRPX2, ATP6AP2, ARHGEF6, NXF5, ZCCHC12, ZNF41, and ZNF81), in which truncating variants or previously published mutations are observed at a relatively high frequency within this cohort. We also highlight 15 other genes (CCDC22, CLIC2, CNKSR2, FRMPD4, HCFC1, IGBP1, KIAA2022, KLF8, MAOA, NAA10, NLGN3, RPL10, SHROOM4, ZDHHC15, and ZNF261) for which replication studies are warranted. We propose that similar reassessment of reported mutations (and genes) with the use of data from large-scale human exome sequencing would be relevant for a wide range of other genetic diseases. PMID:23871722

  15. Genome-scale identification of Legionella pneumophila effectors using a machine learning approach.

    PubMed

    Burstein, David; Zusman, Tal; Degtyar, Elena; Viner, Ram; Segal, Gil; Pupko, Tal

    2009-07-01

    A large number of highly pathogenic bacteria utilize secretion systems to translocate effector proteins into host cells. Using these effectors, the bacteria subvert host cell processes during infection. Legionella pneumophila translocates effectors via the Icm/Dot type-IV secretion system and to date, approximately 100 effectors have been identified by various experimental and computational techniques. Effector identification is a critical first step towards the understanding of the pathogenesis system in L. pneumophila as well as in other bacterial pathogens. Here, we formulate the task of effector identification as a classification problem: each L. pneumophila open reading frame (ORF) was classified as either effector or not. We computationally defined a set of features that best distinguish effectors from non-effectors. These features cover a wide range of characteristics including taxonomical dispersion, regulatory data, genomic organization, similarity to eukaryotic proteomes and more. Machine learning algorithms utilizing these features were then applied to classify all the ORFs within the L. pneumophila genome. Using this approach we were able to predict and experimentally validate 40 new effectors, reaching a success rate of above 90%. Increasing the number of validated effectors to around 140, we were able to gain novel insights into their characteristics. Effectors were found to have low G+C content, supporting the hypothesis that a large number of effectors originate via horizontal gene transfer, probably from their protozoan host. In addition, effectors were found to cluster in specific genomic regions. Finally, we were able to provide a novel description of the C-terminal translocation signal required for effector translocation by the Icm/Dot secretion system. To conclude, we have discovered 40 novel L. pneumophila effectors, predicted over a hundred additional highly probable effectors, and shown the applicability of machine learning algorithms for the identification and characterization of bacterial pathogenesis determinants.

  16. Munitions related feature extraction from LIDAR data.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Roberts, Barry L.

    2010-06-01

    The characterization of former military munitions ranges is critical in the identification of areas likely to contain residual unexploded ordnance (UXO). Although these ranges are large, often covering tens-of-thousands of acres, the actual target areas represent only a small fraction of the sites. The challenge is that many of these sites do not have records indicating locations of former target areas. The identification of target areas is critical in the characterization and remediation of these sites. The Strategic Environmental Research and Development Program (SERDP) and Environmental Security Technology Certification Program (ESTCP) of the DoD have been developing and implementing techniquesmore » for the efficient characterization of large munitions ranges. As part of this process, high-resolution LIDAR terrain data sets have been collected over several former ranges. These data sets have been shown to contain information relating to former munitions usage at these ranges, specifically terrain cratering due to high-explosives detonations. The location and relative intensity of crater features can provide information critical in reconstructing the usage history of a range, and indicate areas most likely to contain UXO. We have developed an automated procedure using an adaptation of the Circular Hough Transform for the identification of crater features in LIDAR terrain data. The Circular Hough Transform is highly adept at finding circular features (craters) in noisy terrain data sets. This technique has the ability to find features of a specific radius providing a means of filtering features based on expected scale and providing additional spatial characterization of the identified feature. This method of automated crater identification has been applied to several former munitions ranges with positive results.« less

  17. A knowledge-based approach to identification and adaptation in dynamical systems control

    NASA Technical Reports Server (NTRS)

    Glass, B. J.; Wong, C. M.

    1988-01-01

    Artificial intelligence techniques are applied to the problems of model form and parameter identification of large-scale dynamic systems. The object-oriented knowledge representation is discussed in the context of causal modeling and qualitative reasoning. Structured sets of rules are used for implementing qualitative component simulations, for catching qualitative discrepancies and quantitative bound violations, and for making reconfiguration and control decisions that affect the physical system. These decisions are executed by backward-chaining through a knowledge base of control action tasks. This approach was implemented for two examples: a triple quadrupole mass spectrometer and a two-phase thermal testbed. Results of tests with both of these systems demonstrate that the software replicates some or most of the functionality of a human operator, thereby reducing the need for a human-in-the-loop in the lower levels of control of these complex systems.

  18. Hybrid propulsion technology program

    NASA Technical Reports Server (NTRS)

    1990-01-01

    Technology was identified which will enable application of hybrid propulsion to manned and unmanned space launch vehicles. Two design concepts are proposed. The first is a hybrid propulsion system using the classical method of regression (classical hybrid) resulting from the flow of oxidizer across a fuel grain surface. The second system uses a self-sustaining gas generator (gas generator hybrid) to produce a fuel rich exhaust that was mixed with oxidizer in a separate combustor. Both systems offer cost and reliability improvement over the existing solid rocket booster and proposed liquid boosters. The designs were evaluated using life cycle cost and reliability. The program consisted of: (1) identification and evaluation of candidate oxidizers and fuels; (2) preliminary evaluation of booster design concepts; (3) preparation of a detailed point design including life cycle costs and reliability analyses; (4) identification of those hybrid specific technologies needing improvement; and (5) preperation of a technology acquisition plan and large scale demonstration plan.

  19. MapReduce implementation of a hybrid spectral library-database search method for large-scale peptide identification.

    PubMed

    Kalyanaraman, Ananth; Cannon, William R; Latt, Benjamin; Baxter, Douglas J

    2011-11-01

    A MapReduce-based implementation called MR-MSPolygraph for parallelizing peptide identification from mass spectrometry data is presented. The underlying serial method, MSPolygraph, uses a novel hybrid approach to match an experimental spectrum against a combination of a protein sequence database and a spectral library. Our MapReduce implementation can run on any Hadoop cluster environment. Experimental results demonstrate that, relative to the serial version, MR-MSPolygraph reduces the time to solution from weeks to hours, for processing tens of thousands of experimental spectra. Speedup and other related performance studies are also reported on a 400-core Hadoop cluster using spectral datasets from environmental microbial communities as inputs. The source code along with user documentation are available on http://compbio.eecs.wsu.edu/MR-MSPolygraph. ananth@eecs.wsu.edu; william.cannon@pnnl.gov. Supplementary data are available at Bioinformatics online.

  20. Dereplication of peptidic natural products through database search of mass spectra

    PubMed Central

    Mohimani, Hosein; Gurevich, Alexey; Mikheenko, Alla; Garg, Neha; Nothias, Louis-Felix; Ninomiya, Akihiro; Takada, Kentaro; Dorrestein, Pieter C.; Pevzner, Pavel A.

    2016-01-01

    Peptidic Natural Products (PNPs) are widely used compounds that include many antibiotics and a variety of other bioactive peptides. While recent breakthroughs in PNP discovery raised the challenge of developing new algorithms for their analysis, identification of PNPs via database search of tandem mass spectra remains an open problem. To address this problem, natural product researchers utilize dereplication strategies that identify known PNPs and lead to the discovery of new ones even in cases when the reference spectra are not present in existing spectral libraries. DEREPLICATOR is a new dereplication algorithm that enabled high-throughput PNP identification and that is compatible with large-scale mass spectrometry-based screening platforms for natural product discovery. After searching nearly one hundred million tandem mass spectra in the Global Natural Products Social (GNPS) molecular networking infrastructure, DEREPLICATOR identified an order of magnitude more PNPs (and their new variants) than any previous dereplication efforts. PMID:27820803

  1. Conservation genetics and genomics of amphibians and reptiles.

    PubMed

    Shaffer, H Bradley; Gidiş, Müge; McCartney-Melstad, Evan; Neal, Kevin M; Oyamaguchi, Hilton M; Tellez, Marisa; Toffelmier, Erin M

    2015-01-01

    Amphibians and reptiles as a group are often secretive, reach their greatest diversity often in remote tropical regions, and contain some of the most endangered groups of organisms on earth. Particularly in the past decade, genetics and genomics have been instrumental in the conservation biology of these cryptic vertebrates, enabling work ranging from the identification of populations subject to trade and exploitation, to the identification of cryptic lineages harboring critical genetic variation, to the analysis of genes controlling key life history traits. In this review, we highlight some of the most important ways that genetic analyses have brought new insights to the conservation of amphibians and reptiles. Although genomics has only recently emerged as part of this conservation tool kit, several large-scale data sources, including full genomes, expressed sequence tags, and transcriptomes, are providing new opportunities to identify key genes, quantify landscape effects, and manage captive breeding stocks of at-risk species.

  2. Aerodynamic coefficient identification package dynamic data accuracy determinations: Lessons learned

    NASA Technical Reports Server (NTRS)

    Heck, M. L.; Findlay, J. T.; Compton, H. R.

    1983-01-01

    The errors in the dynamic data output from the Aerodynamic Coefficient Identification Packages (ACIP) flown on Shuttle flights 1, 3, 4, and 5 were determined using the output from the Inertial Measurement Units (IMU). A weighted least-squares batch algorithm was empolyed. Using an averaging technique, signal detection was enhanced; this allowed improved calibration solutions. Global errors as large as 0.04 deg/sec for the ACIP gyros, 30 mg for linear accelerometers, and 0.5 deg/sec squared in the angular accelerometer channels were detected and removed with a combination is bias, scale factor, misalignment, and g-sensitive calibration constants. No attempt was made to minimize local ACIP dynamic data deviations representing sensed high-frequency vibration or instrument noise. Resulting 1sigma calibrated ACIP global accuracies were within 0.003 eg/sec, 1.0 mg, and 0.05 deg/sec squared for the gyros, linear accelerometers, and angular accelerometers, respectively.

  3. Venus: radar determination of gravity potential.

    PubMed

    Shapiro, I I; Pettengill, G H; Sherman, G N; Rogers, A E; Ingalls, R P

    1973-02-02

    We describe a method for the determination of the gravity potential of Venus from multiple-frequency radar measurements. The method is based on the strong frequency dependence of the absorption of radio waves in Venus' atmosphere. Comparison of the differing radar reflection intensities at several frequencies yields the height of the surface relative to a reference pressure contour; combination with measurements of round-trip echo delays allows the pressure, and hence the gravity potential contour, to be mapped relative to the mean planet radius. Since calibration data from other frequencies are unavailable, the absorption-sensitive Haystack Observatory data have been analyzed under the assumption of uniform surface reflectivity to yield a gravity equipotential contour for the equatorial region and a tentative upper bound of 6 x 10(-4) on the fractional difference of Venus' principal equatorial moments of inertia. The minima in the equipotential contours appear to be associated with topographic minima.

  4. Development of a Compact Eleven Feed Cryostat for the Patriot 12-m Antenna System

    NASA Technical Reports Server (NTRS)

    Beaudoin, Christopher; Kildal, Per-Simon; Yang, Jian; Pantaleev, Miroslav

    2010-01-01

    The Eleven antenna has constant beam width, constant phase center location, and low spillover over a decade bandwidth. Therefore, it can feed a reflector for high aperture efficiency (also called feed efficiency). It is equally important that the feed efficiency and its subefficiencies not be degraded significantly by installing the feed in a cryostat. The MIT Haystack Observatory, with guidance from Onsala Space Observatory and Chalmers University, has been working to integrate the Eleven antenna into a compact cryostat suitable for the Patriot 12-m antenna. Since the analysis of the feed efficiencies in this presentation is purely computational, we first demonstrate the validity of the computed results by comparing them to measurements. Subsequently, we analyze the dependence of the cryostat size on the feed efficiencies, and, lastly, the Patriot 12-m subreflector is incorporated into the computational model to assess the overall broadband efficiency of the antenna system.

  5. Citizen Science

    NASA Technical Reports Server (NTRS)

    Memarsadeghi, Nargess

    2015-01-01

    Scientists and engineers constantly face new challenges, despite myriad advances in computing. More sets of data are collected today from earth and sky than there is time or resources available to carefully analyze them. Some problems either don't have fast algorithms to solve them or have solutions that must be found among millions of options, a situation akin to finding a needle in a haystack. But all hope is not lost: advances in technology and the Internet have empowered the general public to participate in the scientific process via individual computational resources and brain cognition, which isn't matched by any machine. Citizen scientists are volunteers who perform scientific work by making observations, collecting and disseminating data, making measurements, and analyzing or interpreting data without necessarily having any scientific training. In so doing, individuals from all over the world can contribute to science in ways that wouldn't have been otherwise possible.

  6. Detection of protein-small molecule binding using a self-referencing external cavity laser biosensor.

    PubMed

    Meng Zhang; Peh, Jessie; Hergenrother, Paul J; Cunningham, Brian T

    2014-01-01

    High throughput screening of protein-small molecule binding interactions using label-free optical biosensors is challenging, as the detected signals are often similar in magnitude to experimental noise. Here, we describe a novel self-referencing external cavity laser (ECL) biosensor approach that achieves high resolution and high sensitivity, while eliminating thermal noise with sub-picometer wavelength accuracy. Using the self-referencing ECL biosensor, we demonstrate detection of binding between small molecules and a variety of immobilized protein targets with binding affinities or inhibition constants in the sub-nanomolar to low micromolar range. The demonstrated ability to perform detection in the presence of several interfering compounds opens the potential for increasing the throughput of the approach. As an example application, we performed a "needle-in-the-haystack" screen for inhibitors against carbonic anhydrase isozyme II (CA II), in which known inhibitors are clearly differentiated from inactive molecules within a compound library.

  7. Laboratory investigation of nitrile ices of Titan's stratospheric clouds

    NASA Astrophysics Data System (ADS)

    Nna Mvondo, D.; Anderson, C. M.; McLain, J. L.; Samuelson, R. E.

    2017-09-01

    Titan's mid to lower stratosphere contains complex cloud systems of numerous organic ice particles comprised of both hydrocarbon and nitrile compounds. Most of these stratospheric ice clouds form as a result of vapor condensation formation processes. However, there are additional ice emission features such as dicyanoacetylene (C4N2) and the 220 cm-1 ice emission feature (the "Haystack") that are difficult to explain since there are no observed vapor emission features associated with these ices. In our laboratory, using a high-vacuum chamber coupled to a FTIR spectrometer, we are engaged in a dedicated investigation of Titan's stratospheric ices to interpret and constrain Cassini Composite InfraRed Spectrometer (CIRS) far-IR data. We will present laboratory transmittance spectra obtained for propionitrile (CH3CH2CN), cyanogen (C2N2) and hydrogen cyanide (HCN) ices, as well as various combinations of their mixtures, to better understand the cloud chemistry occurring in Titan's stratosphere.

  8. Holism, health and data - managing the person-centred digital haystack.

    PubMed

    Rigby, Michael

    2010-01-01

    There is currently very legitimate pressure to change the function of health information systems to more explicitly serve and support the individual. Concurrently other support services to health, including condition monitoring, social care and intelligent housing, add to the complexity of patient health related data. A paradigm shift is needed, to recognize that many agencies and systems outside as well as within the health sector all work to support the citizen's health, yet the essential integrated view is not being provided, even though much needed not least by the citizen and their agents. Broker technologies can enable this new paradigm. Moreover, a move to identifying the patient's appointments and support services would enable a daily life dimension to be given due weight, and would recognize the patient's viewpoint in wanting to live an ordered life, not one subservient and reactive to the delivery of care by providers.

  9. Implications of very long baseline interferometry measurements on North American intra-plate crustal deformation

    NASA Technical Reports Server (NTRS)

    Allenby, R. J.

    1979-01-01

    Very Long Baseline Interferometry experiments over the last 1-3/4 years between Owens Valley, CA and Haystack, MA Radio Observatories suggest an upper limit of east-west crustal deformation between the two sites of about 1 cm/yr. In view of the fact that the baseline between the two sites traverses most of the major geological provinces of the United States, this low rate of crustal deformation has direct relevance to intra-plate crustal tectonics. The most active region traversed by this baseline is the Basin and Range province, which was estimated by various researchers to be expanding in an east-west direction at rates of .3 to 1.5 cm/yr. The Colorado Plateau and Rocky Mountain system also appear to be expanding, but at a somewhat lower rate, while east of the Rocky Mountains, the predominant stress appears to be compressional, nearly horizontal, and east to northeast trending.

  10. Animal-Friendly Affinity Reagents: Replacing the Needless in the Haystack.

    PubMed

    Gray, A C; Sidhu, S S; Chandrasekera, P C; Hendriksen, C F M; Borrebaeck, C A K

    2016-12-01

    The multibillion-dollar global antibody industry produces an indispensable resource but that is generated using millions of animals. Despite the irrefutable maturation and availability of animal-friendly affinity reagents (AFAs) employing naïve B lymphocyte or synthetic recombinant technologies expressed by phage display, animal immunisation is still authorised for antibody production. Remarkably, replacement opportunities have been overlooked, despite the enormous potential reduction in animal use. Directive 2010/63/EU requires that animals are not used where alternatives exist. To ensure its implementation, we have engaged in discussions with the EU Reference Laboratory for alternatives to animal testing (EURL ECVAM) and the Directorate General for Environment to carve out an EU-led replacement strategy. Measures must be imposed to avoid outsourcing, regulate commercial production, and ensure that antibody producers are fully supported. Copyright © 2016 The Authors. Published by Elsevier Ltd.. All rights reserved.

  11. Zebrafish Whole-Adult-Organism Chemogenomics for Large-Scale Predictive and Discovery Chemical Biology

    PubMed Central

    Lam, Siew Hong; Mathavan, Sinnakarupan; Tong, Yan; Li, Haixia; Karuturi, R. Krishna Murthy; Wu, Yilian; Vega, Vinsensius B.; Liu, Edison T.; Gong, Zhiyuan

    2008-01-01

    The ability to perform large-scale, expression-based chemogenomics on whole adult organisms, as in invertebrate models (worm and fly), is highly desirable for a vertebrate model but its feasibility and potential has not been demonstrated. We performed expression-based chemogenomics on the whole adult organism of a vertebrate model, the zebrafish, and demonstrated its potential for large-scale predictive and discovery chemical biology. Focusing on two classes of compounds with wide implications to human health, polycyclic (halogenated) aromatic hydrocarbons [P(H)AHs] and estrogenic compounds (ECs), we generated robust prediction models that can discriminate compounds of the same class from those of different classes in two large independent experiments. The robust expression signatures led to the identification of biomarkers for potent aryl hydrocarbon receptor (AHR) and estrogen receptor (ER) agonists, respectively, and were validated in multiple targeted tissues. Knowledge-based data mining of human homologs of zebrafish genes revealed highly conserved chemical-induced biological responses/effects, health risks, and novel biological insights associated with AHR and ER that could be inferred to humans. Thus, our study presents an effective, high-throughput strategy of capturing molecular snapshots of chemical-induced biological states of a whole adult vertebrate that provides information on biomarkers of effects, deregulated signaling pathways, and possible affected biological functions, perturbed physiological systems, and increased health risks. These findings place zebrafish in a strategic position to bridge the wide gap between cell-based and rodent models in chemogenomics research and applications, especially in preclinical drug discovery and toxicology. PMID:18618001

  12. Reduced-order model for underwater target identification using proper orthogonal decomposition

    NASA Astrophysics Data System (ADS)

    Ramesh, Sai Sudha; Lim, Kian Meng

    2017-03-01

    Research on underwater acoustics has seen major development over the past decade due to its widespread applications in domains such as underwater communication/navigation (SONAR), seismic exploration and oceanography. In particular, acoustic signatures from partially or fully buried targets can be used in the identification of buried mines for mine counter measures (MCM). Although there exist several techniques to identify target properties based on SONAR images and acoustic signatures, these methods first employ a feature extraction method to represent the dominant characteristics of a data set, followed by the use of an appropriate classifier based on neural networks or the relevance vector machine. The aim of the present study is to demonstrate the applications of proper orthogonal decomposition (POD) technique in capturing dominant features of a set of scattered pressure signals, and subsequent use of the POD modes and coefficients in the identification of partially buried underwater target parameters such as its location, size and material density. Several numerical examples are presented to demonstrate the performance of the system identification method based on POD. Although the present study is based on 2D acoustic model, the method can be easily extended to 3D models and thereby enables cost-effective representations of large-scale data.

  13. Synthesizing spatiotemporally sparse smartphone sensor data for bridge modal identification

    NASA Astrophysics Data System (ADS)

    Ozer, Ekin; Feng, Maria Q.

    2016-08-01

    Smartphones as vibration measurement instruments form a large-scale, citizen-induced, and mobile wireless sensor network (WSN) for system identification and structural health monitoring (SHM) applications. Crowdsourcing-based SHM is possible with a decentralized system granting citizens with operational responsibility and control. Yet, citizen initiatives introduce device mobility, drastically changing SHM results due to uncertainties in the time and the space domains. This paper proposes a modal identification strategy that fuses spatiotemporally sparse SHM data collected by smartphone-based WSNs. Multichannel data sampled with the time and the space independence is used to compose the modal identification parameters such as frequencies and mode shapes. Structural response time history can be gathered by smartphone accelerometers and converted into Fourier spectra by the processor units. Timestamp, data length, energy to power conversion address temporal variation, whereas spatial uncertainties are reduced by geolocation services or determining node identity via QR code labels. Then, parameters collected from each distributed network component can be extended to global behavior to deduce modal parameters without the need of a centralized and synchronous data acquisition system. The proposed method is tested on a pedestrian bridge and compared with a conventional reference monitoring system. The results show that the spatiotemporally sparse mobile WSN data can be used to infer modal parameters despite non-overlapping sensor operation schedule.

  14. Molecular taxonomy and identification within the Antarctic genus Trematomus (Notothenioidei, Teleostei): How valuable is barcoding with COI?

    NASA Astrophysics Data System (ADS)

    Lautredou, A.-C.; Bonillo, C.; Denys, G.; Cruaud, C.; Ozouf-Costaz, C.; Lecointre, G.; Dettai, A.

    2010-08-01

    The Trematominae are a particularly interesting subfamily within the antarctic suborder Notothenioidei (Teleostei). The 14 closely related species occupy a large range of ecological of niches, extremely useful for evolutionary and biogeography studies in the Antarctic Ocean. But some Trematomus species can be difficult to identify by using morphological criteria, specially young stages and damaged specimens. Molecular identification would therefore be highly useful, however the suitability of the cytochrome oxidase I gene in a barcoding approach needs to be assessed. We evaluated species delineation within the genus Trematomus comparing morphological identification, nuclear markers (the rhodopsin retrogene and a new nuclear marker pkd1: polycystic kidney disease 1) and COI. We show that Trematomus vicarius is not distinguishable from Trematomus bernacchii with the molecular markers used, and neither is Trematomus loennbergii from Trematomus lepidorhinus. We suggest that until this is investigated further, studies including these species list them as T. loennbergii/ T. lepidorhinus group, and keep voucher samples and specimens. Generally, COI gives a congruent result with the rhodopsin retrogene, and except for the previously cited species pairs, COI barcoding is efficient for identification in this group. Moreover pkd1 might not be suitable for a phylogenetic study at this scale for this group.

  15. Climatic and physiographic controls on catchment-scale nitrate loss at different spatial scales: insights from a top-down model development approach

    NASA Astrophysics Data System (ADS)

    Shafii, Mahyar; Basu, Nandita; Schiff, Sherry; Van Cappellen, Philippe

    2017-04-01

    Dramatic increase in nitrogen circulating in the biosphere due to anthropogenic activities has resulted in impairment of water quality in groundwater and surface water causing eutrophication in coastal regions. Understanding the fate and transport of nitrogen from landscape to coastal areas requires exploring the drivers of nitrogen processes in both time and space, as well as the identification of appropriate flow pathways. Conceptual models can be used as diagnostic tools to provide insights into such controls. However, diagnostic evaluation of coupled hydrological-biogeochemical models is challenging. This research proposes a top-down methodology utilizing hydrochemical signatures to develop conceptual models for simulating the integrated streamflow and nitrate responses while taking into account dominant controls on nitrate variability (e.g., climate, soil water content, etc.). Our main objective is to seek appropriate model complexity that sufficiently reproduces multiple hydrological and nitrate signatures. Having developed a suitable conceptual model for a given watershed, we employ it in sensitivity studies to demonstrate the dominant process controls that contribute to the nitrate response at scales of interest. We apply the proposed approach to nitrate simulation in a range of small to large sub-watersheds in the Grand River Watershed (GRW) located in Ontario. Such multi-basin modeling experiment will enable us to address process scaling and investigate the consequences of lumping processes in terms of models' predictive capability. The proposed methodology can be applied to the development of large-scale models that can help decision-making associated with nutrients management at regional scale.

  16. Statistical Analyses of Scatterplots to Identify Important Factors in Large-Scale Simulations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kleijnen, J.P.C.; Helton, J.C.

    1999-04-01

    The robustness of procedures for identifying patterns in scatterplots generated in Monte Carlo sensitivity analyses is investigated. These procedures are based on attempts to detect increasingly complex patterns in the scatterplots under consideration and involve the identification of (1) linear relationships with correlation coefficients, (2) monotonic relationships with rank correlation coefficients, (3) trends in central tendency as defined by means, medians and the Kruskal-Wallis statistic, (4) trends in variability as defined by variances and interquartile ranges, and (5) deviations from randomness as defined by the chi-square statistic. The following two topics related to the robustness of these procedures are consideredmore » for a sequence of example analyses with a large model for two-phase fluid flow: the presence of Type I and Type II errors, and the stability of results obtained with independent Latin hypercube samples. Observations from analysis include: (1) Type I errors are unavoidable, (2) Type II errors can occur when inappropriate analysis procedures are used, (3) physical explanations should always be sought for why statistical procedures identify variables as being important, and (4) the identification of important variables tends to be stable for independent Latin hypercube samples.« less

  17. Improvement of proteolytic efficiency towards low-level proteins by an antifouling surface of alumina gel in a microchannel.

    PubMed

    Liu, Yun; Wang, Huixiang; Liu, Qingping; Qu, Haiyun; Liu, Baohong; Yang, Pengyuan

    2010-11-07

    A microfluidic reactor has been developed for rapid enhancement of protein digestion by constructing an alumina network within a poly(ethylene terephthalate) (PET) microchannel. Trypsin is stably immobilized in a sol-gel network on the PET channel surface after pretreatment, which produces a protein-resistant interface to reduce memory effects, as characterized by X-ray fluorescence spectrometry and electroosmotic flow. The gel-derived network within a microchannel provides a large surface-to-volume ratio stationary phase for highly efficient proteolysis of proteins existing both at a low level and in complex extracts. The maximum reaction rate of the encapsulated trypsin reactor, measured by kinetic analysis, is much faster than in bulk solution. Due to the microscopic confinement effect, high levels of enzyme entrapment and the biocompatible microenvironment provided by the alumina gel network, the low-level proteins can be efficiently digested using such a microreactor within a very short residence time of a few seconds. The on-chip microreactor is further applied to the identification of a mixture of proteins extracted from normal mouse liver cytoplasm sample via integration with 2D-LC-ESI-MS/MS to show its potential application for large-scale protein identification.

  18. Free-decay time-domain modal identification for large space structures

    NASA Technical Reports Server (NTRS)

    Kim, Hyoung M.; Vanhorn, David A.; Doiron, Harold H.

    1992-01-01

    Concept definition studies for the Modal Identification Experiment (MIE), a proposed space flight experiment for the Space Station Freedom (SSF), have demonstrated advantages and compatibility of free-decay time-domain modal identification techniques with the on-orbit operational constraints of large space structures. Since practical experience with modal identification using actual free-decay responses of large space structures is very limited, several numerical and test data reduction studies were conducted. Major issues and solutions were addressed, including closely-spaced modes, wide frequency range of interest, data acquisition errors, sampling delay, excitation limitations, nonlinearities, and unknown disturbances during free-decay data acquisition. The data processing strategies developed in these studies were applied to numerical simulations of the MIE, test data from a deployable truss, and launch vehicle flight data. Results of these studies indicate free-decay time-domain modal identification methods can provide accurate modal parameters necessary to characterize the structural dynamics of large space structures.

  19. Teacher Report versus Adaptive Behavior Scale in Assessment of Mental Retardation.

    ERIC Educational Resources Information Center

    Al-Ansari, Ahmed

    1993-01-01

    This study assessed the degree of agreement between teacher report and an adapted Adaptive Behavior Scale in the identification of mental retardation and associated learning difficulties in 257 young Bahraini school children. Findings indicated that the instrument is sensitive in identification of children with mental retardation and exhibits high…

  20. National Identification of Dutch Youth: An Exploratory Study

    ERIC Educational Resources Information Center

    Oppenheimer, Louis

    2011-01-01

    246 Dutch participants aged 8, 10, 12, 14, and 16 years were presented with the Strength of Identification Scale (SoIS; Barrett, 2007) and the National Identity scale based on Cultural and Historical achievements (NICH; derived from the NATID, Keillor & Hult, 1999). The study aimed to examine the extent and nature of Dutch children and…

  1. Genome-wide SNP identification and QTL mapping for black rot resistance in cabbage.

    PubMed

    Lee, Jonghoon; Izzah, Nur Kholilatul; Jayakodi, Murukarthick; Perumal, Sampath; Joh, Ho Jun; Lee, Hyeon Ju; Lee, Sang-Choon; Park, Jee Young; Yang, Ki-Woung; Nou, Il-Sup; Seo, Joodeok; Yoo, Jaeheung; Suh, Youngdeok; Ahn, Kyounggu; Lee, Ji Hyun; Choi, Gyung Ja; Yu, Yeisoo; Kim, Heebal; Yang, Tae-Jin

    2015-02-03

    Black rot is a destructive bacterial disease causing large yield and quality losses in Brassica oleracea. To detect quantitative trait loci (QTL) for black rot resistance, we performed whole-genome resequencing of two cabbage parental lines and genome-wide SNP identification using the recently published B. oleracea genome sequences as reference. Approximately 11.5 Gb of sequencing data was produced from each parental line. Reference genome-guided mapping and SNP calling revealed 674,521 SNPs between the two cabbage lines, with an average of one SNP per 662.5 bp. Among 167 dCAPS markers derived from candidate SNPs, 117 (70.1%) were validated as bona fide SNPs showing polymorphism between the parental lines. We then improved the resolution of a previous genetic map by adding 103 markers including 87 SNP-based dCAPS markers. The new map composed of 368 markers and covers 1467.3 cM with an average interval of 3.88 cM between adjacent markers. We evaluated black rot resistance in the mapping population in three independent inoculation tests using F2:3 progenies and identified one major QTL and three minor QTLs. We report successful utilization of whole-genome resequencing for large-scale SNP identification and development of molecular markers for genetic map construction. In addition, we identified novel QTLs for black rot resistance. The high-density genetic map will promote QTL analysis for other important agricultural traits and marker-assisted breeding of B. oleracea.

  2. Network-assisted target identification for haploinsufficiency and homozygous profiling screens

    PubMed Central

    Wang, Sheng

    2017-01-01

    Chemical genomic screens have recently emerged as a systematic approach to drug discovery on a genome-wide scale. Drug target identification and elucidation of the mechanism of action (MoA) of hits from these noisy high-throughput screens remain difficult. Here, we present GIT (Genetic Interaction Network-Assisted Target Identification), a network analysis method for drug target identification in haploinsufficiency profiling (HIP) and homozygous profiling (HOP) screens. With the drug-induced phenotypic fitness defect of the deletion of a gene, GIT also incorporates the fitness defects of the gene’s neighbors in the genetic interaction network. On three genome-scale yeast chemical genomic screens, GIT substantially outperforms previous scoring methods on target identification on HIP and HOP assays, respectively. Finally, we showed that by combining HIP and HOP assays, GIT further boosts target identification and reveals potential drug’s mechanism of action. PMID:28574983

  3. Deep JVLA Imaging of GOODS-N at 20 cm

    NASA Astrophysics Data System (ADS)

    Owen, Frazer N.

    2018-04-01

    New wideband continuum observations in the 1–2 GHz band of the GOODS-N field using NSF’s Karl G. Jansky Very Large Array (VLA) are presented. The best image with an effective frequency of 1525 MHz reaches an rms noise in the field center of 2.2 μJy, with 1.″6 resolution. A catalog of 795 sources is presented covering a radius of 9 arcminutes centered near the nominal center for the GOODS-N field, very near the nominal VLA pointing center for the observations. Optical/NIR identifications and redshift estimates both from ground-based and HST observations are discussed. Using these optical/NIR data, it is most likely that fewer than 2% of the sources without confusion problems do not have a correct identification. A large subset of the detected sources have radio sizes >1″. It is shown that the radio orientations for such sources correlate well with the HST source orientations, especially for z < 1. This suggests that a least a large subset of the 10 kpc-scale disks of luminous infrared/ultraluminous infrared galaxies (LIRG/ULIRG) have strong star formation, not just in the nucleus. For the half of the objects with z > 1, the sample must be some mixture of very high star formation rates, typically 300 M ⊙ yr‑1, assuming pure star formation, and an active galactic nucleus (AGN) or a mixed AGN/star formation population.

  4. Quantifying the sensitivity of ephemeral streams to land disturbance activities in arid ecosystems at the watershed scale.

    PubMed

    O'Connor, Ben L; Hamada, Yuki; Bowen, Esther E; Grippo, Mark A; Hartmann, Heidi M; Patton, Terri L; Van Lonkhuyzen, Robert A; Carr, Adrianne E

    2014-11-01

    Large areas of public lands administered by the Bureau of Land Management and located in arid regions of the southwestern United States are being considered for the development of utility-scale solar energy facilities. Land-disturbing activities in these desert, alluvium-filled valleys have the potential to adversely affect the hydrologic and ecologic functions of ephemeral streams. Regulation and management of ephemeral streams typically falls under a spectrum of federal, state, and local programs, but scientifically based guidelines for protecting ephemeral streams with respect to land-development activities are largely nonexistent. This study developed an assessment approach for quantifying the sensitivity to land disturbance of ephemeral stream reaches located in proposed solar energy zones (SEZs). The ephemeral stream assessment approach used publicly-available geospatial data on hydrology, topography, surficial geology, and soil characteristics, as well as high-resolution aerial imagery. These datasets were used to inform a professional judgment-based score index of potential land disturbance impacts on selected critical functions of ephemeral streams, including flow and sediment conveyance, ecological habitat value, and groundwater recharge. The total sensitivity scores (sum of scores for the critical stream functions of flow and sediment conveyance, ecological habitats, and groundwater recharge) were used to identify highly sensitive stream reaches to inform decisions on developable areas in SEZs. Total sensitivity scores typically reflected the scores of the individual stream functions; some exceptions pertain to groundwater recharge and ecological habitats. The primary limitations of this assessment approach were the lack of high-resolution identification of ephemeral stream channels in the existing National Hydrography Dataset, and the lack of mechanistic processes describing potential impacts on ephemeral stream functions at the watershed scale. The primary strength of this assessment approach is that it allows watershed-scale planning for low-impact development in arid ecosystems; the qualitative scoring of potential impacts can also be adjusted to accommodate new geospatial data, and to allow for expert and stakeholder input into decisions regarding the identification and potential avoidance of highly sensitive stream reaches.

  5. Quantifying the sensitivity of ephemeral streams to land disturbance activities in arid ecosystems at the watershed scale

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    O’Connor, Ben L.; Hamada, Yuki; Bowen, Esther E.

    2014-08-17

    Large areas of public lands administered by the Bureau of Land Management and located in arid regions of the southwestern United States are being considered for the development of utility-scale solar energy facilities. Land-disturbing activities in these desert, alluvium-filled valleys have the potential to adversely affect the hydrologic and ecologic functions of ephemeral streams. Regulation and management of ephemeral streams typically falls under a spectrum of federal, state, and local programs, but scientifically based guidelines for protecting ephemeral streams with respect to land-development activities are largely nonexistent. This study developed an assessment approach for quantifying the sensitivity to land disturbancemore » of ephemeral stream reaches located in proposed solar energy zones (SEZs). The ephemeral stream assessment approach used publicly-available geospatial data on hydrology, topography, surficial geology, and soil characteristics, as well as highresolution aerial imagery. These datasets were used to inform a professional judgment-based score index of potential land disturbance impacts on selected critical functions of ephemeral streams, including flow and sediment conveyance, ecological habitat value, and groundwater recharge. The total sensitivity scores (sum of scores for the critical stream functions of flow and sediment conveyance, ecological habitats, and groundwater recharge) were used to identify highly sensitive stream reaches to inform decisions on developable areas in SEZs. Total sensitivity scores typically reflected the scores of the individual stream functions; some exceptions pertain to groundwater recharge and ecological habitats. The primary limitations of this assessment approach were the lack of high-resolution identification of ephemeral stream channels in the existing National Hydrography Dataset, and the lack of mechanistic processes describing potential impacts on ephemeral stream functions at the watershed scale.The primary strength of this assessment approach is that it allows watershed-scale planning for low-impact development in arid ecosystems; the qualitative scoring of potential impacts can also be adjusted to accommodate new geospatial data, and to allow for expert and stakeholder input into decisions regarding the identification and potential avoidance of highly sensitive stream reaches.« less

  6. A review of output-only structural mode identification literature employing blind source separation methods

    NASA Astrophysics Data System (ADS)

    Sadhu, A.; Narasimhan, S.; Antoni, J.

    2017-09-01

    Output-only modal identification has seen significant activity in recent years, especially in large-scale structures where controlled input force generation is often difficult to achieve. This has led to the development of new system identification methods which do not require controlled input. They often work satisfactorily if they satisfy some general assumptions - not overly restrictive - regarding the stochasticity of the input. Hundreds of papers covering a wide range of applications appear every year related to the extraction of modal properties from output measurement data in more than two dozen mechanical, aerospace and civil engineering journals. In little more than a decade, concepts of blind source separation (BSS) from the field of acoustic signal processing have been adopted by several researchers and shown that they can be attractive tools to undertake output-only modal identification. Originally intended to separate distinct audio sources from a mixture of recordings, mathematical equivalence to problems in linear structural dynamics have since been firmly established. This has enabled many of the developments in the field of BSS to be modified and applied to output-only modal identification problems. This paper reviews over hundred articles related to the application of BSS and their variants to output-only modal identification. The main contribution of the paper is to present a literature review of the papers which have appeared on the subject. While a brief treatment of the basic ideas are presented where relevant, a comprehensive and critical explanation of their contents is not attempted. Specific issues related to output-only modal identification and the relative advantages and limitations of BSS methods both from theoretical and application standpoints are discussed. Gap areas requiring additional work are also summarized and the paper concludes with possible future trends in this area.

  7. Ambulance Clinical Triage for Acute Stroke Treatment: Paramedic Triage Algorithm for Large Vessel Occlusion.

    PubMed

    Zhao, Henry; Pesavento, Lauren; Coote, Skye; Rodrigues, Edrich; Salvaris, Patrick; Smith, Karen; Bernard, Stephen; Stephenson, Michael; Churilov, Leonid; Yassi, Nawaf; Davis, Stephen M; Campbell, Bruce C V

    2018-04-01

    Clinical triage scales for prehospital recognition of large vessel occlusion (LVO) are limited by low specificity when applied by paramedics. We created the 3-step ambulance clinical triage for acute stroke treatment (ACT-FAST) as the first algorithmic LVO identification tool, designed to improve specificity by recognizing only severe clinical syndromes and optimizing paramedic usability and reliability. The ACT-FAST algorithm consists of (1) unilateral arm drift to stretcher <10 seconds, (2) severe language deficit (if right arm is weak) or gaze deviation/hemineglect assessed by simple shoulder tap test (if left arm is weak), and (3) eligibility and stroke mimic screen. ACT-FAST examination steps were retrospectively validated, and then prospectively validated by paramedics transporting culturally and linguistically diverse patients with suspected stroke in the emergency department, for the identification of internal carotid or proximal middle cerebral artery occlusion. The diagnostic performance of the full ACT-FAST algorithm was then validated for patients accepted for thrombectomy. In retrospective (n=565) and prospective paramedic (n=104) validation, ACT-FAST displayed higher overall accuracy and specificity, when compared with existing LVO triage scales. Agreement of ACT-FAST between paramedics and doctors was excellent (κ=0.91; 95% confidence interval, 0.79-1.0). The full ACT-FAST algorithm (n=60) assessed by paramedics showed high overall accuracy (91.7%), sensitivity (85.7%), specificity (93.5%), and positive predictive value (80%) for recognition of endovascular-eligible LVO. The 3-step ACT-FAST algorithm shows higher specificity and reliability than existing scales for clinical LVO recognition, despite requiring just 2 examination steps. The inclusion of an eligibility step allowed recognition of endovascular-eligible patients with high accuracy. Using a sequential algorithmic approach eliminates scoring confusion and reduces assessment time. Future studies will test whether field application of ACT-FAST by paramedics to bypass suspected patients with LVO directly to endovascular-capable centers can reduce delays to endovascular thrombectomy. © 2018 American Heart Association, Inc.

  8. High-throughput screening and small animal models, where are we?

    PubMed Central

    Giacomotto, Jean; Ségalat, Laurent

    2010-01-01

    Current high-throughput screening methods for drug discovery rely on the existence of targets. Moreover, most of the hits generated during screenings turn out to be invalid after further testing in animal models. To by-pass these limitations, efforts are now being made to screen chemical libraries on whole animals. One of the most commonly used animal model in biology is the murine model Mus musculus. However, its cost limit its use in large-scale therapeutic screening. In contrast, the nematode Caenorhabditis elegans, the fruit fly Drosophila melanogaster, and the fish Danio rerio are gaining momentum as screening tools. These organisms combine genetic amenability, low cost and culture conditions that are compatible with large-scale screens. Their main advantage is to allow high-throughput screening in a whole-animal context. Moreover, their use is not dependent on the prior identification of a target and permits the selection of compounds with an improved safety profile. This review surveys the versatility of these animal models for drug discovery and discuss the options available at this day. PMID:20423335

  9. Effects of selected bacterial cultures on safety and sensory traits of Nocellara Etnea olives produced at large factory scale.

    PubMed

    Randazzo, Cinzia L; Russo, Nunziatina; Pino, Alessandra; Mazzaglia, Agata; Ferrante, Margherita; Conti, Gea Oliveri; Caggia, Cinzia

    2018-05-01

    This work investigates the effects of different combinations of selected lactic acid bacteria strains on Lactobacillus species occurrence, on safety and on sensory traits of natural green table olives, produced at large factory scale. Olives belonging to Nocellara Etnea cv were processed in a 6% NaCl brine and inoculated with six different bacterial cultures, using selected strains belonging to Lactobacillus plantarum, Lactobacillus paracasei and Lactobacillus pentosus species. The fermentation process was strongly influenced by the added starters and the identification of lactic acid bacteria isolated throughout the process confirms that L. pentosus dominated all fermentations, followed by L. plantarum, whereas L. casei was never detected. Pathogens were never found, while histamine and tyrosine were detected in control and in two experimental samples. The samples with the lowest final pH values showed a safer profile and the most appreciated sensory traits. The present study highlights that selected starters promote prevalence of L. pentosus over the autochthonous microbiota throughout the whole process of Nocellara Etnea olives. Copyright © 2018. Published by Elsevier Ltd.

  10. Use of controlled dynamic impacts on hierarchically structured seismically hazardous faults for seismically safe relaxation of shear stresses

    NASA Astrophysics Data System (ADS)

    Ruzhich, Valery V.; Psakhie, Sergey G.; Levina, Elena A.; Shilko, Evgeny V.; Grigoriev, Alexandr S.

    2017-12-01

    In the paper we briefly outline the experience in forecasting catastrophic earthquakes and the general problems in ensuring seismic safety. The purpose of our long-term research is the development and improvement of the methods of man-caused impacts on large-scale fault segments to safely reduce the negative effect of seismodynamic failure. Various laboratory and large-scale field experiments were carried out in the segments of tectonic faults in Baikal rift zone and in main cracks in block-structured ice cove of Lake Baikal using the developed measuring systems and special software for identification and treatment of deformation response of faulty segments to man-caused impacts. The results of the study let us to ground the necessity of development of servo-controlled technologies, which are able to provide changing the shear resistance and deformation regime of fault zone segments by applying vibrational and pulse triggering impacts. We suppose that the use of triggering impacts in highly stressed segments of active faults will promote transferring the geodynamic state of these segments from a metastable to a more stable and safe state.

  11. Recent advances on biological production of difructose dianhydride III.

    PubMed

    Zhu, Yingying; Yu, Shuhuai; Zhang, Wenli; Zhang, Tao; Guang, Cuie; Mu, Wanmeng

    2018-04-01

    Difructose dianhydride III (DFA III) is a cyclic difructose containing two reciprocal glycosidic linkages. It is easily generated with a small amount by sucrose caramelization and thus occurs in a wide range of food-stuffs during food processing. DFA III has half sweetness but only 1/15 energy of sucrose, showing potential industrial application as low-calorie sucrose substitute. In addition, it displays many benefits including prebiotic effect, low cariogenicity property, and hypocholesterolemic effect, and improves absorption of minerals, flavonoids, and immunoglobulin G. DFA III is biologically produced from inulin by inulin fructotransferase (IFTase, EC 4.2.2.18). Plenty of DFA III-producing enzymes have been identified. The crystal structure of inulin fructotransferase has been determined, and its molecular modification has been performed to improve the catalytic activity and structural stability. Large-scale production of DFA III has been studied by various IFTases, especially using an ultrafiltration membrane bioreactor. In this article, the recent findings on physiological effects of DFA III are briefly summarized; the research progresses on identification, expression, and molecular modification of IFTase and large-scale biological production of DFA III by IFTase are reviewed in detail.

  12. STRIDE: Species Tree Root Inference from Gene Duplication Events.

    PubMed

    Emms, David M; Kelly, Steven

    2017-12-01

    The correct interpretation of any phylogenetic tree is dependent on that tree being correctly rooted. We present STRIDE, a fast, effective, and outgroup-free method for identification of gene duplication events and species tree root inference in large-scale molecular phylogenetic analyses. STRIDE identifies sets of well-supported in-group gene duplication events from a set of unrooted gene trees, and analyses these events to infer a probability distribution over an unrooted species tree for the location of its root. We show that STRIDE correctly identifies the root of the species tree in multiple large-scale molecular phylogenetic data sets spanning a wide range of timescales and taxonomic groups. We demonstrate that the novel probability model implemented in STRIDE can accurately represent the ambiguity in species tree root assignment for data sets where information is limited. Furthermore, application of STRIDE to outgroup-free inference of the origin of the eukaryotic tree resulted in a root probability distribution that provides additional support for leading hypotheses for the origin of the eukaryotes. © The Author 2017. Published by Oxford University Press on behalf of the Society for Molecular Biology and Evolution.

  13. Discovering and understanding oncogenic gene fusions through data intensive computational approaches

    PubMed Central

    Latysheva, Natasha S.; Babu, M. Madan

    2016-01-01

    Abstract Although gene fusions have been recognized as important drivers of cancer for decades, our understanding of the prevalence and function of gene fusions has been revolutionized by the rise of next-generation sequencing, advances in bioinformatics theory and an increasing capacity for large-scale computational biology. The computational work on gene fusions has been vastly diverse, and the present state of the literature is fragmented. It will be fruitful to merge three camps of gene fusion bioinformatics that appear to rarely cross over: (i) data-intensive computational work characterizing the molecular biology of gene fusions; (ii) development research on fusion detection tools, candidate fusion prioritization algorithms and dedicated fusion databases and (iii) clinical research that seeks to either therapeutically target fusion transcripts and proteins or leverages advances in detection tools to perform large-scale surveys of gene fusion landscapes in specific cancer types. In this review, we unify these different—yet highly complementary and symbiotic—approaches with the view that increased synergy will catalyze advancements in gene fusion identification, characterization and significance evaluation. PMID:27105842

  14. Task-driven dictionary learning.

    PubMed

    Mairal, Julien; Bach, Francis; Ponce, Jean

    2012-04-01

    Modeling data with linear combinations of a few elements from a learned dictionary has been the focus of much recent research in machine learning, neuroscience, and signal processing. For signals such as natural images that admit such sparse representations, it is now well established that these models are well suited to restoration tasks. In this context, learning the dictionary amounts to solving a large-scale matrix factorization problem, which can be done efficiently with classical optimization tools. The same approach has also been used for learning features from data for other purposes, e.g., image classification, but tuning the dictionary in a supervised way for these tasks has proven to be more difficult. In this paper, we present a general formulation for supervised dictionary learning adapted to a wide variety of tasks, and present an efficient algorithm for solving the corresponding optimization problem. Experiments on handwritten digit classification, digital art identification, nonlinear inverse image problems, and compressed sensing demonstrate that our approach is effective in large-scale settings, and is well suited to supervised and semi-supervised classification, as well as regression tasks for data that admit sparse representations.

  15. Feasibility study on conducting overflight measurements of shaped sonic boom signatures using the Firebee BQM-34E RPV

    NASA Technical Reports Server (NTRS)

    Maglieri, Domenic J.; Sothcott, Victor E.; Keefer, Thomas N., Jr.

    1993-01-01

    A study was performed to determine the feasibility of establishing if a 'shaped' sonic boom signature, experimentally shown in wind tunnel models out to about 10 body lengths, will persist out to representative flight conditions of 200 to 300 body lengths. The study focuses on the use of a relatively large supersonic remotely-piloted and recoverable vehicle. Other simulation methods that may accomplish the objective are also addressed and include the use of nonrecoverable target drones, missiles, full-scale drones, very large wind tunnels, ballistic facilities, whirling-arm techniques, rocket sled tracks, and airplane nose probes. In addition, this report will also present a background on the origin of the feasibility study including a brief review of the equivalent body concept, a listing of the basic sonic boom signature characteristics and requirements, identification of candidate vehicles in terms of desirable features/availability, and vehicle characteristics including geometries, area distributions, and resulting sonic boom signatures. A program is developed that includes wind tunnel sonic boom and force models and tests for both a basic and modified vehicles and full-scale flight tests.

  16. Large-scale detection of antigen-specific T cells using peptide-MHC-I multimers labeled with DNA barcodes.

    PubMed

    Bentzen, Amalie Kai; Marquard, Andrea Marion; Lyngaa, Rikke; Saini, Sunil Kumar; Ramskov, Sofie; Donia, Marco; Such, Lina; Furness, Andrew J S; McGranahan, Nicholas; Rosenthal, Rachel; Straten, Per Thor; Szallasi, Zoltan; Svane, Inge Marie; Swanton, Charles; Quezada, Sergio A; Jakobsen, Søren Nyboe; Eklund, Aron Charles; Hadrup, Sine Reker

    2016-10-01

    Identification of the peptides recognized by individual T cells is important for understanding and treating immune-related diseases. Current cytometry-based approaches are limited to the simultaneous screening of 10-100 distinct T-cell specificities in one sample. Here we use peptide-major histocompatibility complex (MHC) multimers labeled with individual DNA barcodes to screen >1,000 peptide specificities in a single sample, and detect low-frequency CD8 T cells specific for virus- or cancer-restricted antigens. When analyzing T-cell recognition of shared melanoma antigens before and after adoptive cell therapy in melanoma patients, we observe a greater number of melanoma-specific T-cell populations compared with cytometry-based approaches. Furthermore, we detect neoepitope-specific T cells in tumor-infiltrating lymphocytes and peripheral blood from patients with non-small cell lung cancer. Barcode-labeled pMHC multimers enable the combination of functional T-cell analysis with large-scale epitope recognition profiling for the characterization of T-cell recognition in various diseases, including in small clinical samples.

  17. Globalization and human cooperation

    PubMed Central

    Buchan, Nancy R.; Grimalda, Gianluca; Wilson, Rick; Brewer, Marilynn; Fatas, Enrique; Foddy, Margaret

    2009-01-01

    Globalization magnifies the problems that affect all people and that require large-scale human cooperation, for example, the overharvesting of natural resources and human-induced global warming. However, what does globalization imply for the cooperation needed to address such global social dilemmas? Two competing hypotheses are offered. One hypothesis is that globalization prompts reactionary movements that reinforce parochial distinctions among people. Large-scale cooperation then focuses on favoring one's own ethnic, racial, or language group. The alternative hypothesis suggests that globalization strengthens cosmopolitan attitudes by weakening the relevance of ethnicity, locality, or nationhood as sources of identification. In essence, globalization, the increasing interconnectedness of people worldwide, broadens the group boundaries within which individuals perceive they belong. We test these hypotheses by measuring globalization at both the country and individual levels and analyzing the relationship between globalization and individual cooperation with distal others in multilevel sequential cooperation experiments in which players can contribute to individual, local, and/or global accounts. Our samples were drawn from the general populations of the United States, Italy, Russia, Argentina, South Africa, and Iran. We find that as country and individual levels of globalization increase, so too does individual cooperation at the global level vis-à-vis the local level. In essence, “globalized” individuals draw broader group boundaries than others, eschewing parochial motivations in favor of cosmopolitan ones. Globalization may thus be fundamental in shaping contemporary large-scale cooperation and may be a positive force toward the provision of global public goods. PMID:19255433

  18. The XChemExplorer graphical workflow tool for routine or large-scale protein-ligand structure determination.

    PubMed

    Krojer, Tobias; Talon, Romain; Pearce, Nicholas; Collins, Patrick; Douangamath, Alice; Brandao-Neto, Jose; Dias, Alexandre; Marsden, Brian; von Delft, Frank

    2017-03-01

    XChemExplorer (XCE) is a data-management and workflow tool to support large-scale simultaneous analysis of protein-ligand complexes during structure-based ligand discovery (SBLD). The user interfaces of established crystallographic software packages such as CCP4 [Winn et al. (2011), Acta Cryst. D67, 235-242] or PHENIX [Adams et al. (2010), Acta Cryst. D66, 213-221] have entrenched the paradigm that a `project' is concerned with solving one structure. This does not hold for SBLD, where many almost identical structures need to be solved and analysed quickly in one batch of work. Functionality to track progress and annotate structures is essential. XCE provides an intuitive graphical user interface which guides the user from data processing, initial map calculation, ligand identification and refinement up until data dissemination. It provides multiple entry points depending on the need of each project, enables batch processing of multiple data sets and records metadata, progress and annotations in an SQLite database. XCE is freely available and works on any Linux and Mac OS X system, and the only dependency is to have the latest version of CCP4 installed. The design and usage of this tool are described here, and its usefulness is demonstrated in the context of fragment-screening campaigns at the Diamond Light Source. It is routinely used to analyse projects comprising 1000 data sets or more, and therefore scales well to even very large ligand-design projects.

  19. Environmental performance evaluation of large-scale municipal solid waste incinerators using data envelopment analysis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chen, H.-W.; Chang, N.-B., E-mail: nchang@mail.ucf.ed; Chen, J.-C.

    2010-07-15

    Limited to insufficient land resources, incinerators are considered in many countries such as Japan and Germany as the major technology for a waste management scheme capable of dealing with the increasing demand for municipal and industrial solid waste treatment in urban regions. The evaluation of these municipal incinerators in terms of secondary pollution potential, cost-effectiveness, and operational efficiency has become a new focus in the highly interdisciplinary area of production economics, systems analysis, and waste management. This paper aims to demonstrate the application of data envelopment analysis (DEA) - a production economics tool - to evaluate performance-based efficiencies of 19more » large-scale municipal incinerators in Taiwan with different operational conditions. A 4-year operational data set from 2002 to 2005 was collected in support of DEA modeling using Monte Carlo simulation to outline the possibility distributions of operational efficiency of these incinerators. Uncertainty analysis using the Monte Carlo simulation provides a balance between simplifications of our analysis and the soundness of capturing the essential random features that complicate solid waste management systems. To cope with future challenges, efforts in the DEA modeling, systems analysis, and prediction of the performance of large-scale municipal solid waste incinerators under normal operation and special conditions were directed toward generating a compromised assessment procedure. Our research findings will eventually lead to the identification of the optimal management strategies for promoting the quality of solid waste incineration, not only in Taiwan, but also elsewhere in the world.« less

  20. Identification of large-scale meteorological patterns associated with extreme precipitation in the US northeast

    NASA Astrophysics Data System (ADS)

    Agel, Laurie; Barlow, Mathew; Feldstein, Steven B.; Gutowski, William J.

    2018-03-01

    Patterns of daily large-scale circulation associated with Northeast US extreme precipitation are identified using both k-means clustering (KMC) and Self-Organizing Maps (SOM) applied to tropopause height. The tropopause height provides a compact representation of the upper-tropospheric potential vorticity, which is closely related to the overall evolution and intensity of weather systems. Extreme precipitation is defined as the top 1% of daily wet-day observations at 35 Northeast stations, 1979-2008. KMC is applied on extreme precipitation days only, while the SOM algorithm is applied to all days in order to place the extreme results into the overall context of patterns for all days. Six tropopause patterns are identified through KMC for extreme day precipitation: a summertime tropopause ridge, a summertime shallow trough/ridge, a summertime shallow eastern US trough, a deeper wintertime eastern US trough, and two versions of a deep cold-weather trough located across the east-central US. Thirty SOM patterns for all days are identified. Results for all days show that 6 SOM patterns account for almost half of the extreme days, although extreme precipitation occurs in all SOM patterns. The same SOM patterns associated with extreme precipitation also routinely produce non-extreme precipitation; however, on extreme precipitation days the troughs, on average, are deeper and the downstream ridges more pronounced. Analysis of other fields associated with the large-scale patterns show various degrees of anomalously strong moisture transport preceding, and upward motion during, extreme precipitation events.

  1. Efficient data management in a large-scale epidemiology research project.

    PubMed

    Meyer, Jens; Ostrzinski, Stefan; Fredrich, Daniel; Havemann, Christoph; Krafczyk, Janina; Hoffmann, Wolfgang

    2012-09-01

    This article describes the concept of a "Central Data Management" (CDM) and its implementation within the large-scale population-based medical research project "Personalized Medicine". The CDM can be summarized as a conjunction of data capturing, data integration, data storage, data refinement, and data transfer. A wide spectrum of reliable "Extract Transform Load" (ETL) software for automatic integration of data as well as "electronic Case Report Forms" (eCRFs) was developed, in order to integrate decentralized and heterogeneously captured data. Due to the high sensitivity of the captured data, high system resource availability, data privacy, data security and quality assurance are of utmost importance. A complex data model was developed and implemented using an Oracle database in high availability cluster mode in order to integrate different types of participant-related data. Intelligent data capturing and storage mechanisms are improving the quality of data. Data privacy is ensured by a multi-layered role/right system for access control and de-identification of identifying data. A well defined backup process prevents data loss. Over the period of one and a half year, the CDM has captured a wide variety of data in the magnitude of approximately 5terabytes without experiencing any critical incidents of system breakdown or loss of data. The aim of this article is to demonstrate one possible way of establishing a Central Data Management in large-scale medical and epidemiological studies. Copyright © 2011 Elsevier Ireland Ltd. All rights reserved.

  2. Globalization and human cooperation.

    PubMed

    Buchan, Nancy R; Grimalda, Gianluca; Wilson, Rick; Brewer, Marilynn; Fatas, Enrique; Foddy, Margaret

    2009-03-17

    Globalization magnifies the problems that affect all people and that require large-scale human cooperation, for example, the overharvesting of natural resources and human-induced global warming. However, what does globalization imply for the cooperation needed to address such global social dilemmas? Two competing hypotheses are offered. One hypothesis is that globalization prompts reactionary movements that reinforce parochial distinctions among people. Large-scale cooperation then focuses on favoring one's own ethnic, racial, or language group. The alternative hypothesis suggests that globalization strengthens cosmopolitan attitudes by weakening the relevance of ethnicity, locality, or nationhood as sources of identification. In essence, globalization, the increasing interconnectedness of people worldwide, broadens the group boundaries within which individuals perceive they belong. We test these hypotheses by measuring globalization at both the country and individual levels and analyzing the relationship between globalization and individual cooperation with distal others in multilevel sequential cooperation experiments in which players can contribute to individual, local, and/or global accounts. Our samples were drawn from the general populations of the United States, Italy, Russia, Argentina, South Africa, and Iran. We find that as country and individual levels of globalization increase, so too does individual cooperation at the global level vis-à-vis the local level. In essence, "globalized" individuals draw broader group boundaries than others, eschewing parochial motivations in favor of cosmopolitan ones. Globalization may thus be fundamental in shaping contemporary large-scale cooperation and may be a positive force toward the provision of global public goods.

  3. The XChemExplorer graphical workflow tool for routine or large-scale protein–ligand structure determination

    PubMed Central

    Krojer, Tobias; Talon, Romain; Pearce, Nicholas; Douangamath, Alice; Brandao-Neto, Jose; Dias, Alexandre; Marsden, Brian

    2017-01-01

    XChemExplorer (XCE) is a data-management and workflow tool to support large-scale simultaneous analysis of protein–ligand complexes during structure-based ligand discovery (SBLD). The user interfaces of established crystallo­graphic software packages such as CCP4 [Winn et al. (2011 ▸), Acta Cryst. D67, 235–242] or PHENIX [Adams et al. (2010 ▸), Acta Cryst. D66, 213–221] have entrenched the paradigm that a ‘project’ is concerned with solving one structure. This does not hold for SBLD, where many almost identical structures need to be solved and analysed quickly in one batch of work. Functionality to track progress and annotate structures is essential. XCE provides an intuitive graphical user interface which guides the user from data processing, initial map calculation, ligand identification and refinement up until data dissemination. It provides multiple entry points depending on the need of each project, enables batch processing of multiple data sets and records metadata, progress and annotations in an SQLite database. XCE is freely available and works on any Linux and Mac OS X system, and the only dependency is to have the latest version of CCP4 installed. The design and usage of this tool are described here, and its usefulness is demonstrated in the context of fragment-screening campaigns at the Diamond Light Source. It is routinely used to analyse projects comprising 1000 data sets or more, and therefore scales well to even very large ligand-design projects. PMID:28291762

  4. Expanding the user base beyond HEP for the Ganga distributed analysis user interface

    NASA Astrophysics Data System (ADS)

    Currie, R.; Egede, U.; Richards, A.; Slater, M.; Williams, M.

    2017-10-01

    This document presents the result of recent developments within Ganga[1] project to support users from new communities outside of HEP. In particular I will examine the case of users from the Large Scale Survey Telescope (LSST) group looking to use resources provided by the UK based GridPP[2][3] DIRAC[4][5] instance. An example use case is work performed with users from the LSST Virtual Organisation (VO) to distribute the workflow used for galaxy shape identification analyses. This work highlighted some LSST specific challenges which could be well solved by common tools within the HEP community. As a result of this work the LSST community was able to take advantage of GridPP[2][3] resources to perform large computing tasks within the UK.

  5. Convergence between biological, behavioural and genetic determinants of obesity.

    PubMed

    Ghosh, Sujoy; Bouchard, Claude

    2017-12-01

    Multiple biological, behavioural and genetic determinants or correlates of obesity have been identified to date. Genome-wide association studies (GWAS) have contributed to the identification of more than 100 obesity-associated genetic variants, but their roles in causal processes leading to obesity remain largely unknown. Most variants are likely to have tissue-specific regulatory roles through joint contributions to biological pathways and networks, through changes in gene expression that influence quantitative traits, or through the regulation of the epigenome. The recent availability of large-scale functional genomics resources provides an opportunity to re-examine obesity GWAS data to begin elucidating the function of genetic variants. Interrogation of knockout mouse phenotype resources provides a further avenue to test for evidence of convergence between genetic variation and biological or behavioural determinants of obesity.

  6. Dynamic Identification for Control of Large Space Structures

    NASA Technical Reports Server (NTRS)

    Ibrahim, S. R.

    1985-01-01

    This is a compilation of reports by the one author on one subject. It consists of the following five journal articles: (1) A Parametric Study of the Ibrahim Time Domain Modal Identification Algorithm; (2) Large Modal Survey Testing Using the Ibrahim Time Domain Identification Technique; (3) Computation of Normal Modes from Identified Complex Modes; (4) Dynamic Modeling of Structural from Measured Complex Modes; and (5) Time Domain Quasi-Linear Identification of Nonlinear Dynamic Systems.

  7. Computer-aided detection of human cone photoreceptor inner segments using multi-scale circular voting

    NASA Astrophysics Data System (ADS)

    Liu, Jianfei; Dubra, Alfredo; Tam, Johnny

    2016-03-01

    Cone photoreceptors are highly specialized cells responsible for the origin of vision in the human eye. Their inner segments can be noninvasively visualized using adaptive optics scanning light ophthalmoscopes (AOSLOs) with nonconfocal split detection capabilities. Monitoring the number of cones can lead to more precise metrics for real-time diagnosis and assessment of disease progression. Cell identification in split detection AOSLO images is hindered by cell regions with heterogeneous intensity arising from shadowing effects and low contrast boundaries due to overlying blood vessels. Here, we present a multi-scale circular voting approach to overcome these challenges through the novel combination of: 1) iterative circular voting to identify candidate cells based on their circular structures, 2) a multi-scale strategy to identify the optimal circular voting response, and 3) clustering to improve robustness while removing false positives. We acquired images from three healthy subjects at various locations on the retina and manually labeled cell locations to create ground-truth for evaluating the detection accuracy. The images span a large range of cell densities. The overall recall, precision, and F1 score were 91±4%, 84±10%, and 87±7% (Mean±SD). Results showed that our method for the identification of cone photoreceptor inner segments performs well even with low contrast cell boundaries and vessel obscuration. These encouraging results demonstrate that the proposed approach can robustly and accurately identify cells in split detection AOSLO images.

  8. Perspectives on the role of mobility, behavior, and time scales in the spread of diseases.

    PubMed

    Castillo-Chavez, Carlos; Bichara, Derdei; Morin, Benjamin R

    2016-12-20

    The dynamics, control, and evolution of communicable and vector-borne diseases are intimately connected to the joint dynamics of epidemiological, behavioral, and mobility processes that operate across multiple spatial, temporal, and organizational scales. The identification of a theoretical explanatory framework that accounts for the pattern regularity exhibited by a large number of host-parasite systems, including those sustained by host-vector epidemiological dynamics, is but one of the challenges facing the coevolving fields of computational, evolutionary, and theoretical epidemiology. Host-parasite epidemiological patterns, including epidemic outbreaks and endemic recurrent dynamics, are characteristic to well-identified regions of the world; the result of processes and constraints such as strain competition, host and vector mobility, and population structure operating over multiple scales in response to recurrent disturbances (like El Niño) and climatological and environmental perturbations over thousands of years. It is therefore important to identify and quantify the processes responsible for observed epidemiological macroscopic patterns: the result of individual interactions in changing social and ecological landscapes. In this perspective, we touch on some of the issues calling for the identification of an encompassing theoretical explanatory framework by identifying some of the limitations of existing theory, in the context of particular epidemiological systems. Fostering the reenergizing of research that aims at disentangling the role of epidemiological and socioeconomic forces on disease dynamics, better understood as complex adaptive systems, is a key aim of this perspective.

  9. Estimating the Efficiency of Phosphopeptide Identification by Tandem Mass Spectrometry

    NASA Astrophysics Data System (ADS)

    Hsu, Chuan-Chih; Xue, Liang; Arrington, Justine V.; Wang, Pengcheng; Paez Paez, Juan Sebastian; Zhou, Yuan; Zhu, Jian-Kang; Tao, W. Andy

    2017-06-01

    Mass spectrometry has played a significant role in the identification of unknown phosphoproteins and sites of phosphorylation in biological samples. Analyses of protein phosphorylation, particularly large scale phosphoproteomic experiments, have recently been enhanced by efficient enrichment, fast and accurate instrumentation, and better software, but challenges remain because of the low stoichiometry of phosphorylation and poor phosphopeptide ionization efficiency and fragmentation due to neutral loss. Phosphoproteomics has become an important dimension in systems biology studies, and it is essential to have efficient analytical tools to cover a broad range of signaling events. To evaluate current mass spectrometric performance, we present here a novel method to estimate the efficiency of phosphopeptide identification by tandem mass spectrometry. Phosphopeptides were directly isolated from whole plant cell extracts, dephosphorylated, and then incubated with one of three purified kinases—casein kinase II, mitogen-activated protein kinase 6, and SNF-related protein kinase 2.6—along with 16O4- and 18O4-ATP separately for in vitro kinase reactions. Phosphopeptides were enriched and analyzed by LC-MS. The phosphopeptide identification rate was estimated by comparing phosphopeptides identified by tandem mass spectrometry with phosphopeptide pairs generated by stable isotope labeled kinase reactions. Overall, we found that current high speed and high accuracy mass spectrometers can only identify 20%-40% of total phosphopeptides primarily due to relatively poor fragmentation, additional modifications, and low abundance, highlighting the urgent need for continuous efforts to improve phosphopeptide identification efficiency. [Figure not available: see fulltext.

  10. Improvement of the Measure of the Network Survival Rate and its Application to a Japanese Business Relations Network

    NASA Astrophysics Data System (ADS)

    Kawamoto, Hirokazu; Takayasu, Hideki; Takayasu, Misako

    We analyze the typical characteristics of the percolation transition of a large-scale complex network, a Japanese business relation network consisting of approximately 600,000 nodes and 4,000,000 links. By utilizing percolation characteristics, we revise the definition of network survival rate that we previously proposed. The new network survival rate has a strong correlation with the old one. The calculation cost is also much smaller and the number of trials decreases from 100,000 to 1,000. Finally, we discuss the identification of robust and fragile regions using this index.

  11. Alluvial Fans on Mars

    NASA Technical Reports Server (NTRS)

    Kraal, E. R.; Moore, J. M.; Howard, A. D.; Asphaug, E. A.

    2005-01-01

    Moore and Howard [1] reported the discovery of large alluvial fans in craters on Mars. Their initial survey from 0-30 S found that these fans clustered in three distinct regions and occurred at around the +1 km MOLA defined Mars datum. However, due to incomplete image coverage, Moore and Howard [1]could not conduct a comprehensive survey. They also recognized, though did not quantitatively address, gravity scaling issues. Here, we briefly discuss the identification of alluvial fans on Mars, then consider the general equations governing the deposition of alluvial fans and hypothesize a method for learning about grain size in alluvial fans on Mars.

  12. Neurosurgical sapphire handheld probe for intraoperative optical diagnostics, laser coagulation and aspiration of malignant brain tissue

    NASA Astrophysics Data System (ADS)

    Shikunova, Irina A.; Zaytsev, Kirill I.; Stryukov, Dmitrii O.; Dubyanskaya, Evgenia N.; Kurlov, Vladimir N.

    2017-07-01

    In this paper, a handheld contact probe based on sapphire shaped crystal was developed for the intraoperative optical diagnosis and aspiration of malignant brain tissue combined with the laser hemostasis. Such a favorable combination of several functions in a single instrument significantly increases its clinical relevance. It makes possible highly-accurate real-time detection and removal of either large-scale malignancies or even separate invasive cancer cells. The proposed neuroprobe was integrated into the clinical neurosurgical workflow for the intraoperative fluorescence identification and removal of malignant tissues of the brain.

  13. Unravelling the hidden ancestry of American admixed populations.

    PubMed

    Montinaro, Francesco; Busby, George B J; Pascali, Vincenzo L; Myers, Simon; Hellenthal, Garrett; Capelli, Cristian

    2015-03-24

    The movement of people into the Americas has brought different populations into contact, and contemporary American genomes are the product of a range of complex admixture events. Here we apply a haplotype-based ancestry identification approach to a large set of genome-wide SNP data from a variety of American, European and African populations to determine the contributions of different ancestral populations to the Americas. Our results provide a fine-scale characterization of the source populations, identify a series of novel, previously unreported contributions from Africa and Europe and highlight geohistorical structure in the ancestry of American admixed populations.

  14. Stable isotope dimethyl labelling for quantitative proteomics and beyond

    PubMed Central

    Hsu, Jue-Liang; Chen, Shu-Hui

    2016-01-01

    Stable-isotope reductive dimethylation, a cost-effective, simple, robust, reliable and easy-to- multiplex labelling method, is widely applied to quantitative proteomics using liquid chromatography-mass spectrometry. This review focuses on biological applications of stable-isotope dimethyl labelling for a large-scale comparative analysis of protein expression and post-translational modifications based on its unique properties of the labelling chemistry. Some other applications of the labelling method for sample preparation and mass spectrometry-based protein identification and characterization are also summarized. This article is part of the themed issue ‘Quantitative mass spectrometry’. PMID:27644970

  15. The formation of giant low surface brightness galaxies

    NASA Technical Reports Server (NTRS)

    Hoffman, Yehuda; Silk, Joseph; Wyse, Rosemary F. G.

    1992-01-01

    It is demonstrated that the initial structure of galaxies can be strongly affected by their large-scale environments. In particular, rare (about 3 sigma) massive galaxies in voids will have normal bulges, but unevolved, extended disks; it is proposed that the low surface brightness objects Malin I and Malin II are prototypes of this class of object. The model predicts that searches for more examples of 'crouching giants' should be fruitful, but that such galaxies do not provide a substantial fraction of mass in the universe. The identification of dwarf galaxies is relatively unaffected by their environment.

  16. Determination of spectral signatures of substances in natural waters

    NASA Technical Reports Server (NTRS)

    Klemas, V.; Philpot, W. D.; Davis, G.

    1978-01-01

    Optical remote sensing of water pollution offers the possibility of fast, large scale coverage at a relatively low cost. The possibility of using the spectral characteristics of the upwelling light from water for the purpose of ocean water quality monitoring was explained. The work was broken into several broad tasks as follows: (1) definition of a remotely measured spectral signature of water, (2) collection of field data and testing of the signature analysis, and (3) the possibility of using LANDSAT data for the identification of substances in water. An attempt to extract spectral signatures of acid waste and sediment was successful.

  17. Detection and reconstruction of large scale flow structures in a river by means of empirical mode decomposition combined with Hilbert transform

    NASA Astrophysics Data System (ADS)

    Franca, Mário J.; Lemmin, Ulrich

    2014-05-01

    The occurrence of large scale flow structures (LSFS) coherently organized throughout the flow depth has been reported in field and laboratory experiments of flows over gravel beds, especially under low relative submergence conditions. In these, the instantaneous velocity is synchronized over the whole vertical profile oscillating at a low frequency above or below the time-averaged value. The detection of large scale coherently organized regions in the flow field is often difficult since it requires detailed simultaneous observations of the flow velocities at several levels. The present research avoids the detection problem by using an Acoustic Doppler Velocity Profiler (ADVP), which permits measuring three-dimensional velocities quasi-simultaneously over the full water column. Empirical mode decomposition (EMD) combined with the application of the Hilbert transform is then applied to the instantaneous velocity data to detect and isolate LSFS. The present research was carried out in a Swiss river with low relative submergence of 2.9, herein defined as h/D50, (where h is the mean flow depth and D50 the bed grain size diameter for which 50% of the grains have smaller diameters). 3D ADVP instantaneous velocity measurements were made on a 3x5 rectangular horizontal grid (x-y). Fifteen velocity profiles were equally spaced in the spanwise direction with a distance of 10 cm, and in the streamwise direction with a distance of 15 cm. The vertical resolution of the measurements is roughly 0.5 cm. A measuring grid covering a 3D control volume was defined. The instantaneous velocity profiles were measured for 3.5 min with a sampling frequency of 26 Hz. Oscillating LSFS are detected and isolated in the instantaneous velocity signal of the 15 measured profiles. Their 3D cycle geometry is reconstructed and investigated through phase averaging based on the identification of the instantaneous signal phase (related to the Hilbert transform) applied to the original raw signal. Results for all the profiles are consistent and indicate clearly the presence of LSFS throughout the flow depth with impact on the three components of the velocity profile and on the bed friction velocity. A high correlation of the movement is found throughout the flow depth, thus corroborating the hypothesis of large-scale coherent motion evolving over the whole water depth. These latter are characterized in terms of period, horizontal scale and geometry. The high spatial and temporal resolution of our ADVP was crucial for obtaining comprehensive results on coherent structures dynamics. EMD combined with the Hilbert transform have previously been successfully applied to geophysical flow studies. Here we show that this method can also be used for the analysis of river dynamics. In particular, we demonstrate that a clean, well-behaved intrinsic mode function can be obtained from a noisy velocity time series that allowed a precise determination of the vertical structure of the coherent structures. The phase unwrapping of the UMR and the identification of the phase related velocity components brings new insight into the flow dynamics Research supported by the Swiss National Science Foundation (2000-063818). KEY WORDS: large scale flow structures (LSFS); gravel-bed rivers; empirical mode decomposition; Hilbert transform

  18. Spatial heterogeneity in the Mediterranean Biodiversity Hotspot affects barcoding accuracy of its freshwater fishes.

    PubMed

    Geiger, M F; Herder, F; Monaghan, M T; Almada, V; Barbieri, R; Bariche, M; Berrebi, P; Bohlen, J; Casal-Lopez, M; Delmastro, G B; Denys, G P J; Dettai, A; Doadrio, I; Kalogianni, E; Kärst, H; Kottelat, M; Kovačić, M; Laporte, M; Lorenzoni, M; Marčić, Z; Özuluğ, M; Perdices, A; Perea, S; Persat, H; Porcelotti, S; Puzzi, C; Robalo, J; Šanda, R; Schneider, M; Šlechtová, V; Stoumboudi, M; Walter, S; Freyhof, J

    2014-11-01

    Incomplete knowledge of biodiversity remains a stumbling block for conservation planning and even occurs within globally important Biodiversity Hotspots (BH). Although technical advances have boosted the power of molecular biodiversity assessments, the link between DNA sequences and species and the analytics to discriminate entities remain crucial. Here, we present an analysis of the first DNA barcode library for the freshwater fish fauna of the Mediterranean BH (526 spp.), with virtually complete species coverage (498 spp., 98% extant species). In order to build an identification system supporting conservation, we compared species determination by taxonomists to multiple clustering analyses of DNA barcodes for 3165 specimens. The congruence of barcode clusters with morphological determination was strongly dependent on the method of cluster delineation, but was highest with the general mixed Yule-coalescent (GMYC) model-based approach (83% of all species recovered as GMYC entity). Overall, genetic morphological discontinuities suggest the existence of up to 64 previously unrecognized candidate species. We found reduced identification accuracy when using the entire DNA-barcode database, compared with analyses on databases for individual river catchments. This scale effect has important implications for barcoding assessments and suggests that fairly simple identification pipelines provide sufficient resolution in local applications. We calculated Evolutionarily Distinct and Globally Endangered scores in order to identify candidate species for conservation priority and argue that the evolutionary content of barcode data can be used to detect priority species for future IUCN assessments. We show that large-scale barcoding inventories of complex biotas are feasible and contribute directly to the evaluation of conservation priorities. © 2014 John Wiley & Sons Ltd.

  19. Identification of Tropical-Extratropical Interactions and Extreme Precipitation Events in the Middle East Based On Potential Vorticity and Moisture Transport

    NASA Astrophysics Data System (ADS)

    de Vries, A. J.; Ouwersloot, H. G.; Feldstein, S. B.; Riemer, M.; El Kenawy, A. M.; McCabe, M. F.; Lelieveld, J.

    2018-01-01

    Extreme precipitation events in the otherwise arid Middle East can cause flooding with dramatic socioeconomic impacts. Most of these events are associated with tropical-extratropical interactions, whereby a stratospheric potential vorticity (PV) intrusion reaches deep into the subtropics and forces an incursion of high poleward vertically integrated water vapor transport (IVT) into the Middle East. This study presents an object-based identification method for extreme precipitation events based on the combination of these two larger-scale meteorological features. The general motivation for this approach is that precipitation is often poorly simulated in relatively coarse weather and climate models, whereas the synoptic-scale circulation is much better represented. The algorithm is applied to ERA-Interim reanalysis data (1979-2015) and detects 90% (83%) of the 99th (97.5th) percentile of extreme precipitation days in the region of interest. Our results show that stratospheric PV intrusions and IVT structures are intimately connected to extreme precipitation intensity and seasonality. The farther south a stratospheric PV intrusion reaches, the larger the IVT magnitude, and the longer the duration of their combined occurrence, the more extreme the precipitation. Our algorithm detects a large fraction of the climatological rainfall amounts (40-70%), heavy precipitation days (50-80%), and the top 10 extreme precipitation days (60-90%) at many sites in southern Israel and the northern and western parts of Saudi Arabia. This identification method provides a new tool for future work to disentangle teleconnections, assess medium-range predictability, and improve understanding of climatic changes of extreme precipitation in the Middle East and elsewhere.

  20. Measuring health-related problem solving among African Americans with multiple chronic conditions: application of Rasch analysis.

    PubMed

    Fitzpatrick, Stephanie L; Hill-Briggs, Felicia

    2015-10-01

    Identification of patients with poor chronic disease self-management skills can facilitate treatment planning, determine effectiveness of interventions, and reduce disease complications. This paper describes the use of a Rasch model, the Rating Scale Model, to examine psychometric properties of the 50-item Health Problem-Solving Scale (HPSS) among 320 African American patients with high risk for cardiovascular disease. Items on the positive/effective HPSS subscales targeted patients at low, moderate, and high levels of positive/effective problem solving, whereas items on the negative/ineffective problem solving subscales mostly targeted those at moderate or high levels of ineffective problem solving. Validity was examined by correlating factor scores on the measure with clinical and behavioral measures. Items on the HPSS show promise in the ability to assess health-related problem solving among high risk patients. However, further revisions of the scale are needed to increase its usability and validity with large, diverse patient populations in the future.

Top