Science.gov

Sample records for model organism database

  1. MOPED: Model Organism Protein Expression Database.

    PubMed

    Kolker, Eugene; Higdon, Roger; Haynes, Winston; Welch, Dean; Broomall, William; Lancet, Doron; Stanberry, Larissa; Kolker, Natali

    2012-01-01

    Large numbers of mass spectrometry proteomics studies are being conducted to understand all types of biological processes. The size and complexity of proteomics data hinders efforts to easily share, integrate, query and compare the studies. The Model Organism Protein Expression Database (MOPED, htttp://moped.proteinspire.org) is a new and expanding proteomics resource that enables rapid browsing of protein expression information from publicly available studies on humans and model organisms. MOPED is designed to simplify the comparison and sharing of proteomics data for the greater research community. MOPED uniquely provides protein level expression data, meta-analysis capabilities and quantitative data from standardized analysis. Data can be queried for specific proteins, browsed based on organism, tissue, localization and condition and sorted by false discovery rate and expression. MOPED empowers users to visualize their own expression data and compare it with existing studies. Further, MOPED links to various protein and pathway databases, including GeneCards, Entrez, UniProt, KEGG and Reactome. The current version of MOPED contains over 43,000 proteins with at least one spectral match and more than 11 million high certainty spectra.

  2. Curation accuracy of model organism databases.

    PubMed

    Keseler, Ingrid M; Skrzypek, Marek; Weerasinghe, Deepika; Chen, Albert Y; Fulcher, Carol; Li, Gene-Wei; Lemmer, Kimberly C; Mladinich, Katherine M; Chow, Edmond D; Sherlock, Gavin; Karp, Peter D

    2014-01-01

    Manual extraction of information from the biomedical literature-or biocuration-is the central methodology used to construct many biological databases. For example, the UniProt protein database, the EcoCyc Escherichia coli database and the Candida Genome Database (CGD) are all based on biocuration. Biological databases are used extensively by life science researchers, as online encyclopedias, as aids in the interpretation of new experimental data and as golden standards for the development of new bioinformatics algorithms. Although manual curation has been assumed to be highly accurate, we are aware of only one previous study of biocuration accuracy. We assessed the accuracy of EcoCyc and CGD by manually selecting curated assertions within randomly chosen EcoCyc and CGD gene pages and by then validating that the data found in the referenced publications supported those assertions. A database assertion is considered to be in error if that assertion could not be found in the publication cited for that assertion. We identified 10 errors in the 633 facts that we validated across the two databases, for an overall error rate of 1.58%, and individual error rates of 1.82% for CGD and 1.40% for EcoCyc. These data suggest that manual curation of the experimental literature by Ph.D-level scientists is highly accurate. Database URL: http://ecocyc.org/, http://www.candidagenome.org//

  3. Xenbase: expansion and updates of the Xenopus model organism database

    PubMed Central

    James-Zorn, Christina; Ponferrada, Virgilio G.; Jarabek, Chris J.; Burns, Kevin A.; Segerdell, Erik J.; Lee, Jacqueline; Snyder, Kevin; Bhattacharyya, Bishnu; Karpinka, J. Brad; Fortriede, Joshua; Bowes, Jeff B.; Zorn, Aaron M.; Vize, Peter D.

    2013-01-01

    Xenbase (http://www.xenbase.org) is a model organism database that provides genomic, molecular, cellular and developmental biology content to biomedical researchers working with the frog, Xenopus and Xenopus data to workers using other model organisms. As an amphibian Xenopus serves as a useful evolutionary bridge between invertebrates and more complex vertebrates such as birds and mammals. Xenbase content is collated from a variety of external sources using automated and semi-automated pipelines then processed via a combination of automated and manual annotation. A link-matching system allows for the wide variety of synonyms used to describe biological data on unique features, such as a gene or an anatomical entity, to be used by the database in an equivalent manner. Recent updates to the database include the Xenopus laevis genome, a new Xenopus tropicalis genome build, epigenomic data, collections of RNA and protein sequences associated with genes, more powerful gene expression searches, a community and curated wiki, an extensive set of manually annotated gene expression patterns and a new database module that contains data on over 700 antibodies that are useful for exploring Xenopus cell and developmental biology. PMID:23125366

  4. EcoCyc: fusing model organism databases with systems biology.

    PubMed

    Keseler, Ingrid M; Mackie, Amanda; Peralta-Gil, Martin; Santos-Zavaleta, Alberto; Gama-Castro, Socorro; Bonavides-Martínez, César; Fulcher, Carol; Huerta, Araceli M; Kothari, Anamika; Krummenacker, Markus; Latendresse, Mario; Muñiz-Rascado, Luis; Ong, Quang; Paley, Suzanne; Schröder, Imke; Shearer, Alexander G; Subhraveti, Pallavi; Travers, Mike; Weerasinghe, Deepika; Weiss, Verena; Collado-Vides, Julio; Gunsalus, Robert P; Paulsen, Ian; Karp, Peter D

    2013-01-01

    EcoCyc (http://EcoCyc.org) is a model organism database built on the genome sequence of Escherichia coli K-12 MG1655. Expert manual curation of the functions of individual E. coli gene products in EcoCyc has been based on information found in the experimental literature for E. coli K-12-derived strains. Updates to EcoCyc content continue to improve the comprehensive picture of E. coli biology. The utility of EcoCyc is enhanced by new tools available on the EcoCyc web site, and the development of EcoCyc as a teaching tool is increasing the impact of the knowledge collected in EcoCyc.

  5. ZFIN, the Zebrafish Model Organism Database: updates and new directions

    PubMed Central

    Ruzicka, Leyla; Bradford, Yvonne M.; Frazer, Ken; Howe, Douglas G.; Paddock, Holly; Ramachandran, Sridhar; Singer, Amy; Toro, Sabrina; Van Slyke, Ceri E.; Eagle, Anne E.; Fashena, David; Kalita, Patrick; Knight, Jonathan; Mani, Prita; Martin, Ryan; Moxon, Sierra A. T.; Pich, Christian; Schaper, Kevin; Shao, Xiang; Westerfield, Monte

    2015-01-01

    The Zebrafish Model Organism Database (ZFIN; http://zfin.org) is the central resource for genetic and genomic data from zebrafish (Danio rerio) research. ZFIN staff curate detailed information about genes, mutants, genotypes, reporter lines, sequences, constructs, antibodies, knockdown reagents, expression patterns, phenotypes, gene product function, and orthology from publications. Researchers can submit mutant, transgenic, expression, and phenotype data directly to ZFIN and use the ZFIN Community Wiki to share antibody and protocol information. Data can be accessed through topic-specific searches, a new site-wide search, and the data-mining resource ZebrafishMine (http://zebrafishmine.org). Data download and web service options are also available. ZFIN collaborates with major bioinformatics organizations to verify and integrate genomic sequence data, provide nomenclature support, establish reciprocal links and participate in the development of standardized structured vocabularies (ontologies) used for data annotation and searching. ZFIN-curated gene, function, expression, and phenotype data are available for comparative exploration at several multi-species resources. The use of zebrafish as a model for human disease is increasing. ZFIN is supporting this growing area with three major projects: adding easy access to computed orthology data from gene pages, curating details of the gene expression pattern changes in mutant fish, and curating zebrafish models of human diseases. PMID:26097180

  6. ZFIN, The zebrafish model organism database: Updates and new directions.

    PubMed

    Ruzicka, Leyla; Bradford, Yvonne M; Frazer, Ken; Howe, Douglas G; Paddock, Holly; Ramachandran, Sridhar; Singer, Amy; Toro, Sabrina; Van Slyke, Ceri E; Eagle, Anne E; Fashena, David; Kalita, Patrick; Knight, Jonathan; Mani, Prita; Martin, Ryan; Moxon, Sierra A T; Pich, Christian; Schaper, Kevin; Shao, Xiang; Westerfield, Monte

    2015-08-01

    The Zebrafish Model Organism Database (ZFIN; http://zfin.org) is the central resource for genetic and genomic data from zebrafish (Danio rerio) research. ZFIN staff curate detailed information about genes, mutants, genotypes, reporter lines, sequences, constructs, antibodies, knockdown reagents, expression patterns, phenotypes, gene product function, and orthology from publications. Researchers can submit mutant, transgenic, expression, and phenotype data directly to ZFIN and use the ZFIN Community Wiki to share antibody and protocol information. Data can be accessed through topic-specific searches, a new site-wide search, and the data-mining resource ZebrafishMine (http://zebrafishmine.org). Data download and web service options are also available. ZFIN collaborates with major bioinformatics organizations to verify and integrate genomic sequence data, provide nomenclature support, establish reciprocal links, and participate in the development of standardized structured vocabularies (ontologies) used for data annotation and searching. ZFIN-curated gene, function, expression, and phenotype data are available for comparative exploration at several multi-species resources. The use of zebrafish as a model for human disease is increasing. ZFIN is supporting this growing area with three major projects: adding easy access to computed orthology data from gene pages, curating details of the gene expression pattern changes in mutant fish, and curating zebrafish models of human diseases. PMID:26097180

  7. ZFIN, The zebrafish model organism database: Updates and new directions.

    PubMed

    Ruzicka, Leyla; Bradford, Yvonne M; Frazer, Ken; Howe, Douglas G; Paddock, Holly; Ramachandran, Sridhar; Singer, Amy; Toro, Sabrina; Van Slyke, Ceri E; Eagle, Anne E; Fashena, David; Kalita, Patrick; Knight, Jonathan; Mani, Prita; Martin, Ryan; Moxon, Sierra A T; Pich, Christian; Schaper, Kevin; Shao, Xiang; Westerfield, Monte

    2015-08-01

    The Zebrafish Model Organism Database (ZFIN; http://zfin.org) is the central resource for genetic and genomic data from zebrafish (Danio rerio) research. ZFIN staff curate detailed information about genes, mutants, genotypes, reporter lines, sequences, constructs, antibodies, knockdown reagents, expression patterns, phenotypes, gene product function, and orthology from publications. Researchers can submit mutant, transgenic, expression, and phenotype data directly to ZFIN and use the ZFIN Community Wiki to share antibody and protocol information. Data can be accessed through topic-specific searches, a new site-wide search, and the data-mining resource ZebrafishMine (http://zebrafishmine.org). Data download and web service options are also available. ZFIN collaborates with major bioinformatics organizations to verify and integrate genomic sequence data, provide nomenclature support, establish reciprocal links, and participate in the development of standardized structured vocabularies (ontologies) used for data annotation and searching. ZFIN-curated gene, function, expression, and phenotype data are available for comparative exploration at several multi-species resources. The use of zebrafish as a model for human disease is increasing. ZFIN is supporting this growing area with three major projects: adding easy access to computed orthology data from gene pages, curating details of the gene expression pattern changes in mutant fish, and curating zebrafish models of human diseases.

  8. Choosing a Genome Browser for a Model Organism Database (MOD): Surveying the Maize Community

    Technology Transfer Automated Retrieval System (TEKTRAN)

    As the maize genome sequencing is nearing its completion, the Maize Genetics and Genomics Database (MaizeGDB), the Model Organism Database for maize, integrated a genome browser to its already existing Web interface and database. The addition of the MaizeGDB Genome Browser to MaizeGDB will allow it ...

  9. Model organism databases: essential resources that need the support of both funders and users.

    PubMed

    Oliver, Stephen G; Lock, Antonia; Harris, Midori A; Nurse, Paul; Wood, Valerie

    2016-01-01

    Modern biomedical research depends critically on access to databases that house and disseminate genetic, genomic, molecular, and cell biological knowledge. Even as the explosion of available genome sequences and associated genome-scale data continues apace, the sustainability of professionally maintained biological databases is under threat due to policy changes by major funding agencies. Here, we focus on model organism databases to demonstrate the myriad ways in which biological databases not only act as repositories but actively facilitate advances in research. We present data that show that reducing financial support to model organism databases could prove to be not just scientifically, but also economically, unsound.

  10. The Zebrafish Information Network (ZFIN): the zebrafish model organism database.

    PubMed

    Sprague, Judy; Clements, Dave; Conlin, Tom; Edwards, Pat; Frazer, Ken; Schaper, Kevin; Segerdell, Erik; Song, Peiran; Sprunger, Brock; Westerfield, Monte

    2003-01-01

    The Zebrafish Information Network (ZFIN) is a web based community resource that serves as a centralized location for the curation and integration of zebrafish genetic, genomic and developmental data. ZFIN is publicly accessible at http://zfin.org. ZFIN provides an integrated representation of mutants, genes, genetic markers, mapping panels, publications and community contact data. Recent enhancements to ZFIN include: (i) an anatomical dictionary that provides a controlled vocabulary of anatomical terms, grouped by developmental stages, that may be used to annotate and query gene expression data; (ii) gene expression data; (iii) expanded support for genome sequence; (iv) gene annotation using the standardized vocabulary of Gene Ontology (GO) terms that can be used to elucidate relationships between gene products in zebrafish and other organisms; and (v) collaborations with other databases (NCBI, Sanger Institute and SWISS-PROT) to provide standardization and interconnections based on shared curation.

  11. MaizeGDB update: New tools, data, and interface for the maize model organism database

    Technology Transfer Automated Retrieval System (TEKTRAN)

    MaizeGDB is a highly curated, community-oriented database and informatics service to researchers focused on the crop plant and model organism Zea mays ssp. mays. Although some form of the maize community database has existed over the last 25 years, there have only been two major releases. In 1991, ...

  12. MaizeGDB, the maize model organism database

    Technology Transfer Automated Retrieval System (TEKTRAN)

    MaizeGDB is the maize research community's database for maize genetic and genomic information. In this seminar I will outline our current endeavors including a full website redesign, the status of maize genome assembly and annotation projects, and work toward genome functional annotation. Mechanis...

  13. Using semantic data modeling techniques to organize an object-oriented database for extending the mass storage model

    NASA Technical Reports Server (NTRS)

    Campbell, William J.; Short, Nicholas M., Jr.; Roelofs, Larry H.; Dorfman, Erik

    1991-01-01

    A methodology for optimizing organization of data obtained by NASA earth and space missions is discussed. The methodology uses a concept based on semantic data modeling techniques implemented in a hierarchical storage model. The modeling is used to organize objects in mass storage devices, relational database systems, and object-oriented databases. The semantic data modeling at the metadata record level is examined, including the simulation of a knowledge base and semantic metadata storage issues. The semantic data model hierarchy and its application for efficient data storage is addressed, as is the mapping of the application structure to the mass storage.

  14. Integrated interactions database: tissue-specific view of the human and model organism interactomes.

    PubMed

    Kotlyar, Max; Pastrello, Chiara; Sheahan, Nicholas; Jurisica, Igor

    2016-01-01

    IID (Integrated Interactions Database) is the first database providing tissue-specific protein-protein interactions (PPIs) for model organisms and human. IID covers six species (S. cerevisiae (yeast), C. elegans (worm), D. melonogaster (fly), R. norvegicus (rat), M. musculus (mouse) and H. sapiens (human)) and up to 30 tissues per species. Users query IID by providing a set of proteins or PPIs from any of these organisms, and specifying species and tissues where IID should search for interactions. If query proteins are not from the selected species, IID enables searches across species and tissues automatically by using their orthologs; for example, retrieving interactions in a given tissue, conserved in human and mouse. Interaction data in IID comprises three types of PPI networks: experimentally detected PPIs from major databases, orthologous PPIs and high-confidence computationally predicted PPIs. Interactions are assigned to tissues where their proteins pairs or encoding genes are expressed. IID is a major replacement of the I2D interaction database, with larger PPI networks (a total of 1,566,043 PPIs among 68,831 proteins), tissue annotations for interactions, and new query, analysis and data visualization capabilities. IID is available at http://ophid.utoronto.ca/iid.

  15. The Generic Genome Browser: A Building Block for a Model Organism System Database

    PubMed Central

    Stein, Lincoln D.; Mungall, Christopher; Shu, ShengQiang; Caudy, Michael; Mangone, Marco; Day, Allen; Nickerson, Elizabeth; Stajich, Jason E.; Harris, Todd W.; Arva, Adrian; Lewis, Suzanna

    2002-01-01

    The Generic Model Organism System Database Project (GMOD) seeks to develop reusable software components for model organism system databases. In this paper we describe the Generic Genome Browser (GBrowse), a Web-based application for displaying genomic annotations and other features. For the end user, features of the browser include the ability to scroll and zoom through arbitrary regions of a genome, to enter a region of the genome by searching for a landmark or performing a full text search of all features, and the ability to enable and disable tracks and change their relative order and appearance. The user can upload private annotations to view them in the context of the public ones, and publish those annotations to the community. For the data provider, features of the browser software include reliance on readily available open source components, simple installation, flexible configuration, and easy integration with other components of a model organism system Web site. GBrowse is freely available under an open source license. The software, its documentation, and support are available at http://www.gmod.org. PMID:12368253

  16. MaizeGDB update: new tools, data and interface for the maize model organism database.

    PubMed

    Andorf, Carson M; Cannon, Ethalinda K; Portwood, John L; Gardiner, Jack M; Harper, Lisa C; Schaeffer, Mary L; Braun, Bremen L; Campbell, Darwin A; Vinnakota, Abhinav G; Sribalusu, Venktanaga V; Huerta, Miranda; Cho, Kyoung Tak; Wimalanathan, Kokulapalan; Richter, Jacqueline D; Mauch, Emily D; Rao, Bhavani S; Birkett, Scott M; Sen, Taner Z; Lawrence-Dill, Carolyn J

    2016-01-01

    MaizeGDB is a highly curated, community-oriented database and informatics service to researchers focused on the crop plant and model organism Zea mays ssp. mays. Although some form of the maize community database has existed over the last 25 years, there have only been two major releases. In 1991, the original maize genetics database MaizeDB was created. In 2003, the combined contents of MaizeDB and the sequence data from ZmDB were made accessible as a single resource named MaizeGDB. Over the next decade, MaizeGDB became more sequence driven while still maintaining traditional maize genetics datasets. This enabled the project to meet the continued growing and evolving needs of the maize research community, yet the interface and underlying infrastructure remained unchanged. In 2015, the MaizeGDB team completed a multi-year effort to update the MaizeGDB resource by reorganizing existing data, upgrading hardware and infrastructure, creating new tools, incorporating new data types (including diversity data, expression data, gene models, and metabolic pathways), and developing and deploying a modern interface. In addition to coordinating a data resource, the MaizeGDB team coordinates activities and provides technical support to the maize research community. MaizeGDB is accessible online at http://www.maizegdb.org.

  17. MaizeGDB update: new tools, data and interface for the maize model organism database.

    PubMed

    Andorf, Carson M; Cannon, Ethalinda K; Portwood, John L; Gardiner, Jack M; Harper, Lisa C; Schaeffer, Mary L; Braun, Bremen L; Campbell, Darwin A; Vinnakota, Abhinav G; Sribalusu, Venktanaga V; Huerta, Miranda; Cho, Kyoung Tak; Wimalanathan, Kokulapalan; Richter, Jacqueline D; Mauch, Emily D; Rao, Bhavani S; Birkett, Scott M; Sen, Taner Z; Lawrence-Dill, Carolyn J

    2016-01-01

    MaizeGDB is a highly curated, community-oriented database and informatics service to researchers focused on the crop plant and model organism Zea mays ssp. mays. Although some form of the maize community database has existed over the last 25 years, there have only been two major releases. In 1991, the original maize genetics database MaizeDB was created. In 2003, the combined contents of MaizeDB and the sequence data from ZmDB were made accessible as a single resource named MaizeGDB. Over the next decade, MaizeGDB became more sequence driven while still maintaining traditional maize genetics datasets. This enabled the project to meet the continued growing and evolving needs of the maize research community, yet the interface and underlying infrastructure remained unchanged. In 2015, the MaizeGDB team completed a multi-year effort to update the MaizeGDB resource by reorganizing existing data, upgrading hardware and infrastructure, creating new tools, incorporating new data types (including diversity data, expression data, gene models, and metabolic pathways), and developing and deploying a modern interface. In addition to coordinating a data resource, the MaizeGDB team coordinates activities and provides technical support to the maize research community. MaizeGDB is accessible online at http://www.maizegdb.org. PMID:26432828

  18. MaizeGDB update: new tools, data and interface for the maize model organism database

    PubMed Central

    Andorf, Carson M.; Cannon, Ethalinda K.; Portwood, John L.; Gardiner, Jack M.; Harper, Lisa C.; Schaeffer, Mary L.; Braun, Bremen L.; Campbell, Darwin A.; Vinnakota, Abhinav G.; Sribalusu, Venktanaga V.; Huerta, Miranda; Cho, Kyoung Tak; Wimalanathan, Kokulapalan; Richter, Jacqueline D.; Mauch, Emily D.; Rao, Bhavani S.; Birkett, Scott M.; Sen, Taner Z.; Lawrence-Dill, Carolyn J.

    2016-01-01

    MaizeGDB is a highly curated, community-oriented database and informatics service to researchers focused on the crop plant and model organism Zea mays ssp. mays. Although some form of the maize community database has existed over the last 25 years, there have only been two major releases. In 1991, the original maize genetics database MaizeDB was created. In 2003, the combined contents of MaizeDB and the sequence data from ZmDB were made accessible as a single resource named MaizeGDB. Over the next decade, MaizeGDB became more sequence driven while still maintaining traditional maize genetics datasets. This enabled the project to meet the continued growing and evolving needs of the maize research community, yet the interface and underlying infrastructure remained unchanged. In 2015, the MaizeGDB team completed a multi-year effort to update the MaizeGDB resource by reorganizing existing data, upgrading hardware and infrastructure, creating new tools, incorporating new data types (including diversity data, expression data, gene models, and metabolic pathways), and developing and deploying a modern interface. In addition to coordinating a data resource, the MaizeGDB team coordinates activities and provides technical support to the maize research community. MaizeGDB is accessible online at http://www.maizegdb.org. PMID:26432828

  19. MyMpn: a database for the systems biology model organism Mycoplasma pneumoniae.

    PubMed

    Wodke, Judith A H; Alibés, Andreu; Cozzuto, Luca; Hermoso, Antonio; Yus, Eva; Lluch-Senar, Maria; Serrano, Luis; Roma, Guglielmo

    2015-01-01

    MyMpn (http://mympn.crg.eu) is an online resource devoted to studying the human pathogen Mycoplasma pneumoniae, a minimal bacterium causing lower respiratory tract infections. Due to its small size, its ability to grow in vitro, and the amount of data produced over the past decades, M. pneumoniae is an interesting model organisms for the development of systems biology approaches for unicellular organisms. Our database hosts a wealth of omics-scale datasets generated by hundreds of experimental and computational analyses. These include data obtained from gene expression profiling experiments, gene essentiality studies, protein abundance profiling, protein complex analysis, metabolic reactions and network modeling, cell growth experiments, comparative genomics and 3D tomography. In addition, the intuitive web interface provides access to several visualization and analysis tools as well as to different data search options. The availability and--even more relevant--the accessibility of properly structured and organized data are of up-most importance when aiming to understand the biology of an organism on a global scale. Therefore, MyMpn constitutes a unique and valuable new resource for the large systems biology and microbiology community.

  20. ZFIN, the Zebrafish Model Organism Database: increased support for mutants and transgenics.

    PubMed

    Howe, Douglas G; Bradford, Yvonne M; Conlin, Tom; Eagle, Anne E; Fashena, David; Frazer, Ken; Knight, Jonathan; Mani, Prita; Martin, Ryan; Moxon, Sierra A Taylor; Paddock, Holly; Pich, Christian; Ramachandran, Sridhar; Ruef, Barbara J; Ruzicka, Leyla; Schaper, Kevin; Shao, Xiang; Singer, Amy; Sprunger, Brock; Van Slyke, Ceri E; Westerfield, Monte

    2013-01-01

    ZFIN, the Zebrafish Model Organism Database (http://zfin.org), is the central resource for zebrafish genetic, genomic, phenotypic and developmental data. ZFIN curators manually curate and integrate comprehensive data involving zebrafish genes, mutants, transgenics, phenotypes, genotypes, gene expressions, morpholinos, antibodies, anatomical structures and publications. Integrated views of these data, as well as data gathered through collaborations and data exchanges, are provided through a wide selection of web-based search forms. Among the vertebrate model organisms, zebrafish are uniquely well suited for rapid and targeted generation of mutant lines. The recent rapid production of mutants and transgenic zebrafish is making management of data associated with these resources particularly important to the research community. Here, we describe recent enhancements to ZFIN aimed at improving our support for mutant and transgenic lines, including (i) enhanced mutant/transgenic search functionality; (ii) more expressive phenotype curation methods; (iii) new downloads files and archival data access; (iv) incorporation of new data loads from laboratories undertaking large-scale generation of mutant or transgenic lines and (v) new GBrowse tracks for transgenic insertions, genes with antibodies and morpholinos.

  1. ZFIN, the Zebrafish Model Organism Database: increased support for mutants and transgenics.

    PubMed

    Howe, Douglas G; Bradford, Yvonne M; Conlin, Tom; Eagle, Anne E; Fashena, David; Frazer, Ken; Knight, Jonathan; Mani, Prita; Martin, Ryan; Moxon, Sierra A Taylor; Paddock, Holly; Pich, Christian; Ramachandran, Sridhar; Ruef, Barbara J; Ruzicka, Leyla; Schaper, Kevin; Shao, Xiang; Singer, Amy; Sprunger, Brock; Van Slyke, Ceri E; Westerfield, Monte

    2013-01-01

    ZFIN, the Zebrafish Model Organism Database (http://zfin.org), is the central resource for zebrafish genetic, genomic, phenotypic and developmental data. ZFIN curators manually curate and integrate comprehensive data involving zebrafish genes, mutants, transgenics, phenotypes, genotypes, gene expressions, morpholinos, antibodies, anatomical structures and publications. Integrated views of these data, as well as data gathered through collaborations and data exchanges, are provided through a wide selection of web-based search forms. Among the vertebrate model organisms, zebrafish are uniquely well suited for rapid and targeted generation of mutant lines. The recent rapid production of mutants and transgenic zebrafish is making management of data associated with these resources particularly important to the research community. Here, we describe recent enhancements to ZFIN aimed at improving our support for mutant and transgenic lines, including (i) enhanced mutant/transgenic search functionality; (ii) more expressive phenotype curation methods; (iii) new downloads files and archival data access; (iv) incorporation of new data loads from laboratories undertaking large-scale generation of mutant or transgenic lines and (v) new GBrowse tracks for transgenic insertions, genes with antibodies and morpholinos. PMID:23074187

  2. Xenbase, the Xenopus model organism database; new virtualized system, data types and genomes.

    PubMed

    Karpinka, J Brad; Fortriede, Joshua D; Burns, Kevin A; James-Zorn, Christina; Ponferrada, Virgilio G; Lee, Jacqueline; Karimi, Kamran; Zorn, Aaron M; Vize, Peter D

    2015-01-01

    Xenbase (http://www.xenbase.org), the Xenopus frog model organism database, integrates a wide variety of data from this biomedical model genus. Two closely related species are represented: the allotetraploid Xenopus laevis that is widely used for microinjection and tissue explant-based protocols, and the diploid Xenopus tropicalis which is used for genetics and gene targeting. The two species are extremely similar and protocols, reagents and results from each species are often interchangeable. Xenbase imports, indexes, curates and manages data from both species; all of which are mapped via unique IDs and can be queried in either a species-specific or species agnostic manner. All our services have now migrated to a private cloud to achieve better performance and reliability. We have added new content, including providing full support for morpholino reagents, used to inhibit mRNA translation or splicing and binding to regulatory microRNAs. New genomes assembled by the JGI for both species and are displayed in Gbrowse and are also available for searches using BLAST. Researchers can easily navigate from genome content to gene page reports, literature, experimental reagents and many other features using hyperlinks. Xenbase has also greatly expanded image content for figures published in papers describing Xenopus research via PubMedCentral.

  3. Xenbase, the Xenopus model organism database; new virtualized system, data types and genomes

    PubMed Central

    Karpinka, J. Brad; Fortriede, Joshua D.; Burns, Kevin A.; James-Zorn, Christina; Ponferrada, Virgilio G.; Lee, Jacqueline; Karimi, Kamran; Zorn, Aaron M.; Vize, Peter D.

    2015-01-01

    Xenbase (http://www.xenbase.org), the Xenopus frog model organism database, integrates a wide variety of data from this biomedical model genus. Two closely related species are represented: the allotetraploid Xenopus laevis that is widely used for microinjection and tissue explant-based protocols, and the diploid Xenopus tropicalis which is used for genetics and gene targeting. The two species are extremely similar and protocols, reagents and results from each species are often interchangeable. Xenbase imports, indexes, curates and manages data from both species; all of which are mapped via unique IDs and can be queried in either a species-specific or species agnostic manner. All our services have now migrated to a private cloud to achieve better performance and reliability. We have added new content, including providing full support for morpholino reagents, used to inhibit mRNA translation or splicing and binding to regulatory microRNAs. New genomes assembled by the JGI for both species and are displayed in Gbrowse and are also available for searches using BLAST. Researchers can easily navigate from genome content to gene page reports, literature, experimental reagents and many other features using hyperlinks. Xenbase has also greatly expanded image content for figures published in papers describing Xenopus research via PubMedCentral. PMID:25313157

  4. Developing a biocuration workflow for AgBase, a non-model organism database

    PubMed Central

    Pillai, Lakshmi; Chouvarine, Philippe; Tudor, Catalina O.; Schmidt, Carl J.; Vijay-Shanker, K.; McCarthy, Fiona M.

    2012-01-01

    AgBase provides annotation for agricultural gene products using the Gene Ontology (GO) and Plant Ontology, as appropriate. Unlike model organism species, agricultural species have a body of literature that does not just focus on gene function; to improve efficiency, we use text mining to identify literature for curation. The first component of our annotation interface is the gene prioritization interface that ranks gene products for annotation. Biocurators select the top-ranked gene and mark annotation for these genes as ‘in progress’ or ‘completed’; links enable biocurators to move directly to our biocuration interface (BI). Our BI includes all current GO annotation for gene products and is the main interface to add/modify AgBase curation data. The BI also displays Extracting Genic Information from Text (eGIFT) results for each gene product. eGIFT is a web-based, text-mining tool that associates ranked, informative terms (iTerms) and the articles and sentences containing them, with genes. Moreover, iTerms are linked to GO terms, where they match either a GO term name or a synonym. This enables AgBase biocurators to rapidly identify literature for further curation based on possible GO terms. Because most agricultural species do not have standardized literature, eGIFT searches all gene names and synonyms to associate articles with genes. As many of the gene names can be ambiguous, eGIFT applies a disambiguation step to remove matches that do not correspond to this gene, and filtering is applied to remove abstracts that mention a gene in passing. The BI is linked to our Journal Database (JDB) where corresponding journal citations are stored. Just as importantly, biocurators also add to the JDB citations that have no GO annotation. The AgBase BI also supports bulk annotation upload to facilitate our Inferred from electronic annotation of agricultural gene products. All annotations must pass standard GO Consortium quality checking before release in Ag

  5. Combining next-generation sequencing and online databases for microsatellite development in non-model organisms

    PubMed Central

    Rico, Ciro; Normandeau, Eric; Dion-Côté, Anne-Marie; Rico, María Inés; Côté, Guillaume; Bernatchez, Louis

    2013-01-01

    Next-generation sequencing (NGS) is revolutionising marker development and the rapidly increasing amount of transcriptomes published across a wide variety of taxa is providing valuable sequence databases for the identification of genetic markers without the need to generate new sequences. Microsatellites are still the most important source of polymorphic markers in ecology and evolution. Motivated by our long-term interest in the adaptive radiation of a non-model species complex of whitefishes (Coregonus spp.), in this study, we focus on microsatellite characterisation and multiplex optimisation using transcriptome sequences generated by Illumina® and Roche-454, as well as online databases of Expressed Sequence Tags (EST) for the study of whitefish evolution and demographic history. We identified and optimised 40 polymorphic loci in multiplex PCR reactions and validated the robustness of our analyses by testing several population genetics and phylogeographic predictions using 494 fish from five lakes and 2 distinct ecotypes. PMID:24296905

  6. Immediate dissemination of student discoveries to a model organism database enhances classroom-based research experiences.

    PubMed

    Wiley, Emily A; Stover, Nicholas A

    2014-01-01

    Use of inquiry-based research modules in the classroom has soared over recent years, largely in response to national calls for teaching that provides experience with scientific processes and methodologies. To increase the visibility of in-class studies among interested researchers and to strengthen their impact on student learning, we have extended the typical model of inquiry-based labs to include a means for targeted dissemination of student-generated discoveries. This initiative required: 1) creating a set of research-based lab activities with the potential to yield results that a particular scientific community would find useful and 2) developing a means for immediate sharing of student-generated results. Working toward these goals, we designed guides for course-based research aimed to fulfill the need for functional annotation of the Tetrahymena thermophila genome, and developed an interactive Web database that links directly to the official Tetrahymena Genome Database for immediate, targeted dissemination of student discoveries. This combination of research via the course modules and the opportunity for students to immediately "publish" their novel results on a Web database actively used by outside scientists culminated in a motivational tool that enhanced students' efforts to engage the scientific process and pursue additional research opportunities beyond the course.

  7. Immediate Dissemination of Student Discoveries to a Model Organism Database Enhances Classroom-Based Research Experiences

    PubMed Central

    Wiley, Emily A.; Stover, Nicholas A.

    2014-01-01

    Use of inquiry-based research modules in the classroom has soared over recent years, largely in response to national calls for teaching that provides experience with scientific processes and methodologies. To increase the visibility of in-class studies among interested researchers and to strengthen their impact on student learning, we have extended the typical model of inquiry-based labs to include a means for targeted dissemination of student-generated discoveries. This initiative required: 1) creating a set of research-based lab activities with the potential to yield results that a particular scientific community would find useful and 2) developing a means for immediate sharing of student-generated results. Working toward these goals, we designed guides for course-based research aimed to fulfill the need for functional annotation of the Tetrahymena thermophila genome, and developed an interactive Web database that links directly to the official Tetrahymena Genome Database for immediate, targeted dissemination of student discoveries. This combination of research via the course modules and the opportunity for students to immediately “publish” their novel results on a Web database actively used by outside scientists culminated in a motivational tool that enhanced students’ efforts to engage the scientific process and pursue additional research opportunities beyond the course. PMID:24591511

  8. A regional estimate of soil organic carbon content linking the RothC model to spatialised climate and soil database

    NASA Astrophysics Data System (ADS)

    Sirca, Costantino; Salis, Michele; Spano, Donatella

    2014-05-01

    Soil organic carbon (SOC) represents the largest pool of organic carbon in the biosphere, and plays a vital role in ecosystem function determining soil fertility, water holding capacity, and susceptibility to land degradation. The SOC amount is mainly led by soil type, land use, and climate. In this work an assessment of SOC pools in Mediterranean soils is presented. The SOC content was estimated at regional scale in Sardinia, the second largest island of the Mediterranean Basin, linking the RothC model (Rothamsted Carbon model) to a high detailed spatialised climate, land use and soil database. More than 300 soil analysis data, covering different land use typologies, were used for the calibration and validation of the model. A good agreement between soil C estimated from the model and ground data was found. The methodology allowed to obtain the current SOC pools estimation for the different land use categories at regional scale.

  9. MaizeGDB: The Maize Model Organism Database for Basic, Translational, and Applied Research

    PubMed Central

    Lawrence, Carolyn J.; Harper, Lisa C.; Schaeffer, Mary L.; Sen, Taner Z.; Seigfried, Trent E.; Campbell, Darwin A.

    2008-01-01

    In 2001 maize became the number one production crop in the world with the Food and Agriculture Organization of the United Nations reporting over 614 million tonnes produced. Its success is due to the high productivity per acre in tandem with a wide variety of commercial uses. Not only is maize an excellent source of food, feed, and fuel, but also its by-products are used in the production of various commercial products. Maize's unparalleled success in agriculture stems from basic research, the outcomes of which drive breeding and product development. In order for basic, translational, and applied researchers to benefit from others' investigations, newly generated data must be made freely and easily accessible. MaizeGDB is the maize research community's central repository for genetics and genomics information. The overall goals of MaizeGDB are to facilitate access to the outcomes of maize research by integrating new maize data into the database and to support the maize research community by coordinating group activities. PMID:18769488

  10. Database for propagation models

    NASA Technical Reports Server (NTRS)

    Kantak, Anil V.

    1991-01-01

    A propagation researcher or a systems engineer who intends to use the results of a propagation experiment is generally faced with various database tasks such as the selection of the computer software, the hardware, and the writing of the programs to pass the data through the models of interest. This task is repeated every time a new experiment is conducted or the same experiment is carried out at a different location generating different data. Thus the users of this data have to spend a considerable portion of their time learning how to implement the computer hardware and the software towards the desired end. This situation may be facilitated considerably if an easily accessible propagation database is created that has all the accepted (standardized) propagation phenomena models approved by the propagation research community. Also, the handling of data will become easier for the user. Such a database construction can only stimulate the growth of the propagation research it if is available to all the researchers, so that the results of the experiment conducted by one researcher can be examined independently by another, without different hardware and software being used. The database may be made flexible so that the researchers need not be confined only to the contents of the database. Another way in which the database may help the researchers is by the fact that they will not have to document the software and hardware tools used in their research since the propagation research community will know the database already. The following sections show a possible database construction, as well as properties of the database for the propagation research.

  11. Use of model organism and disease databases to support matchmaking for human disease gene discovery.

    PubMed

    Mungall, Christopher J; Washington, Nicole L; Nguyen-Xuan, Jeremy; Condit, Christopher; Smedley, Damian; Köhler, Sebastian; Groza, Tudor; Shefchek, Kent; Hochheiser, Harry; Robinson, Peter N; Lewis, Suzanna E; Haendel, Melissa A

    2015-10-01

    The Matchmaker Exchange application programming interface (API) allows searching a patient's genotypic or phenotypic profiles across clinical sites, for the purposes of cohort discovery and variant disease causal validation. This API can be used not only to search for matching patients, but also to match against public disease and model organism data. This public disease data enable matching known diseases and variant-phenotype associations using phenotype semantic similarity algorithms developed by the Monarch Initiative. The model data can provide additional evidence to aid diagnosis, suggest relevant models for disease mechanism and treatment exploration, and identify collaborators across the translational divide. The Monarch Initiative provides an implementation of this API for searching multiple integrated sources of data that contextualize the knowledge about any given patient or patient family into the greater biomedical knowledge landscape. While this corpus of data can aid diagnosis, it is also the beginning of research to improve understanding of rare human diseases.

  12. Choosing a genome browser for a Model Organism Database: surveying the Maize community

    PubMed Central

    Sen, Taner Z.; Harper, Lisa C.; Schaeffer, Mary L.; Andorf, Carson M.; Seigfried, Trent E.; Campbell, Darwin A.; Lawrence, Carolyn J.

    2010-01-01

    As the B73 maize genome sequencing project neared completion, MaizeGDB began to integrate a graphical genome browser with its existing web interface and database. To ensure that maize researchers would optimally benefit from the potential addition of a genome browser to the existing MaizeGDB resource, personnel at MaizeGDB surveyed researchers’ needs. Collected data indicate that existing genome browsers for maize were inadequate and suggest implementation of a browser with quick interface and intuitive tools would meet most researchers’ needs. Here, we document the survey’s outcomes, review functionalities of available genome browser software platforms and offer our rationale for choosing the GBrowse software suite for MaizeGDB. Because the genome as represented within the MaizeGDB Genome Browser is tied to detailed phenotypic data, molecular marker information, available stocks, etc., the MaizeGDB Genome Browser represents a novel mechanism by which the researchers can leverage maize sequence information toward crop improvement directly. Database URL: http://gbrowse.maizegdb.org/ PMID:20627860

  13. Choosing a genome browser for a Model Organism Database: surveying the maize community.

    PubMed

    Sen, Taner Z; Harper, Lisa C; Schaeffer, Mary L; Andorf, Carson M; Seigfried, Trent E; Campbell, Darwin A; Lawrence, Carolyn J

    2010-01-01

    As the B73 maize genome sequencing project neared completion, MaizeGDB began to integrate a graphical genome browser with its existing web interface and database. To ensure that maize researchers would optimally benefit from the potential addition of a genome browser to the existing MaizeGDB resource, personnel at MaizeGDB surveyed researchers' needs. Collected data indicate that existing genome browsers for maize were inadequate and suggest implementation of a browser with quick interface and intuitive tools would meet most researchers' needs. Here, we document the survey's outcomes, review functionalities of available genome browser software platforms and offer our rationale for choosing the GBrowse software suite for MaizeGDB. Because the genome as represented within the MaizeGDB Genome Browser is tied to detailed phenotypic data, molecular marker information, available stocks, etc., the MaizeGDB Genome Browser represents a novel mechanism by which the researchers can leverage maize sequence information toward crop improvement directly. Database URL: http://gbrowse.maizegdb.org/

  14. Integration of an Evidence Base into a Probabilistic Risk Assessment Model. The Integrated Medical Model Database: An Organized Evidence Base for Assessing In-Flight Crew Health Risk and System Design

    NASA Technical Reports Server (NTRS)

    Saile, Lynn; Lopez, Vilma; Bickham, Grandin; FreiredeCarvalho, Mary; Kerstman, Eric; Byrne, Vicky; Butler, Douglas; Myers, Jerry; Walton, Marlei

    2011-01-01

    This slide presentation reviews the Integrated Medical Model (IMM) database, which is an organized evidence base for assessing in-flight crew health risk. The database is a relational database accessible to many people. The database quantifies the model inputs by a ranking based on the highest value of the data as Level of Evidence (LOE) and the quality of evidence (QOE) score that provides an assessment of the evidence base for each medical condition. The IMM evidence base has already been able to provide invaluable information for designers, and for other uses.

  15. Building a Database for a Quantitative Model

    NASA Technical Reports Server (NTRS)

    Kahn, C. Joseph; Kleinhammer, Roger

    2014-01-01

    A database can greatly benefit a quantitative analysis. The defining characteristic of a quantitative risk, or reliability, model is the use of failure estimate data. Models can easily contain a thousand Basic Events, relying on hundreds of individual data sources. Obviously, entering so much data by hand will eventually lead to errors. Not so obviously entering data this way does not aid linking the Basic Events to the data sources. The best way to organize large amounts of data on a computer is with a database. But a model does not require a large, enterprise-level database with dedicated developers and administrators. A database built in Excel can be quite sufficient. A simple spreadsheet database can link every Basic Event to the individual data source selected for them. This database can also contain the manipulations appropriate for how the data is used in the model. These manipulations include stressing factors based on use and maintenance cycles, dormancy, unique failure modes, the modeling of multiple items as a single "Super component" Basic Event, and Bayesian Updating based on flight and testing experience. A simple, unique metadata field in both the model and database provides a link from any Basic Event in the model to its data source and all relevant calculations. The credibility for the entire model often rests on the credibility and traceability of the data.

  16. Organizing a breast cancer database: data management.

    PubMed

    Yi, Min; Hunt, Kelly K

    2016-06-01

    Developing and organizing a breast cancer database can provide data and serve as valuable research tools for those interested in the etiology, diagnosis, and treatment of cancer. Depending on the research setting, the quality of the data can be a major issue. Assuring that the data collection process does not contribute inaccuracies can help to assure the overall quality of subsequent analyses. Data management is work that involves the planning, development, implementation, and administration of systems for the acquisition, storage, and retrieval of data while protecting it by implementing high security levels. A properly designed database provides you with access to up-to-date, accurate information. Database design is an important component of application design. If you take the time to design your databases properly, you'll be rewarded with a solid application foundation on which you can build the rest of your application.

  17. Organizing a breast cancer database: data management.

    PubMed

    Yi, Min; Hunt, Kelly K

    2016-06-01

    Developing and organizing a breast cancer database can provide data and serve as valuable research tools for those interested in the etiology, diagnosis, and treatment of cancer. Depending on the research setting, the quality of the data can be a major issue. Assuring that the data collection process does not contribute inaccuracies can help to assure the overall quality of subsequent analyses. Data management is work that involves the planning, development, implementation, and administration of systems for the acquisition, storage, and retrieval of data while protecting it by implementing high security levels. A properly designed database provides you with access to up-to-date, accurate information. Database design is an important component of application design. If you take the time to design your databases properly, you'll be rewarded with a solid application foundation on which you can build the rest of your application. PMID:27197511

  18. A database for propagation models

    NASA Technical Reports Server (NTRS)

    Kantak, Anil V.; Suwitra, Krisjani; Le, Chuong

    1995-01-01

    A database of various propagation phenomena models that can be used by telecommunications systems engineers to obtain parameter values for systems design is presented. This is an easy-to-use tool and is currently available for either a PC using Excel software under Windows environment or a Macintosh using Excel software for Macintosh. All the steps necessary to use the software are easy and many times self explanatory.

  19. Modeling and Databases for Teaching Petrology

    NASA Astrophysics Data System (ADS)

    Asher, P.; Dutrow, B.

    2003-12-01

    With the widespread availability of high-speed computers with massive storage and ready transport capability of large amounts of data, computational and petrologic modeling and the use of databases provide new tools with which to teach petrology. Modeling can be used to gain insights into a system, predict system behavior, describe a system's processes, compare with a natural system or simply to be illustrative. These aspects result from data driven or empirical, analytical or numerical models or the concurrent examination of multiple lines of evidence. At the same time, use of models can enhance core foundations of the geosciences by improving critical thinking skills and by reinforcing prior knowledge gained. However, the use of modeling to teach petrology is dictated by the level of expectation we have for students and their facility with modeling approaches. For example, do we expect students to push buttons and navigate a program, understand the conceptual model and/or evaluate the results of a model. Whatever the desired level of sophistication, specific elements of design should be incorporated into a modeling exercise for effective teaching. These include, but are not limited to; use of the scientific method, use of prior knowledge, a clear statement of purpose and goals, attainable goals, a connection to the natural/actual system, a demonstration that complex heterogeneous natural systems are amenable to analyses by these techniques and, ideally, connections to other disciplines and the larger earth system. Databases offer another avenue with which to explore petrology. Large datasets are available that allow integration of multiple lines of evidence to attack a petrologic problem or understand a petrologic process. These are collected into a database that offers a tool for exploring, organizing and analyzing the data. For example, datasets may be geochemical, mineralogic, experimental and/or visual in nature, covering global, regional to local scales

  20. Nonbibliographic Databases in a Corporate Health, Safety, and Environment Organization.

    ERIC Educational Resources Information Center

    Cubillas, Mary M.

    1981-01-01

    Summarizes the characteristics of TOXIN, CHEMFILE, and the Product Profile Information System (PPIS), nonbibliographic databases used by Shell Oil Company's Health, Safety, and Environment Organization. (FM)

  1. Software Engineering Laboratory (SEL) database organization and user's guide

    NASA Technical Reports Server (NTRS)

    So, Maria; Heller, Gerard; Steinberg, Sandra; Spiegel, Douglas

    1989-01-01

    The organization of the Software Engineering Laboratory (SEL) database is presented. Included are definitions and detailed descriptions of the database tables and views, the SEL data, and system support data. The mapping from the SEL and system support data to the base tables is described. In addition, techniques for accessing the database, through the Database Access Manager for the SEL (DAMSEL) system and via the ORACLE structured query language (SQL), are discussed.

  2. Community Organizing for Database Trial Buy-In by Patrons

    ERIC Educational Resources Information Center

    Pionke, J. J.

    2015-01-01

    Database trials do not often garner a lot of feedback. Using community-organizing techniques can not only potentially increase the amount of feedback received but also deepen the relationship between the librarian and his or her constituent group. This is a case study of the use of community-organizing techniques in a series of database trials for…

  3. Conceptual and logical level of database modeling

    NASA Astrophysics Data System (ADS)

    Hunka, Frantisek; Matula, Jiri

    2016-06-01

    Conceptual and logical levels form the top most levels of database modeling. Usually, ORM (Object Role Modeling) and ER diagrams are utilized to capture the corresponding schema. The final aim of business process modeling is to store its results in the form of database solution. For this reason, value oriented business process modeling which utilizes ER diagram to express the modeling entities and relationships between them are used. However, ER diagrams form the logical level of database schema. To extend possibilities of different business process modeling methodologies, the conceptual level of database modeling is needed. The paper deals with the REA value modeling approach to business process modeling using ER-diagrams, and derives conceptual model utilizing ORM modeling approach. Conceptual model extends possibilities for value modeling to other business modeling approaches.

  4. DEPOT: A Database of Environmental Parameters, Organizations and Tools

    SciTech Connect

    CARSON,SUSAN D.; HUNTER,REGINA LEE; MALCZYNSKI,LEONARD A.; POHL,PHILLIP I.; QUINTANA,ENRICO; SOUZA,CAROLINE A.; HIGLEY,KATHRYN; MURPHIE,WILLIAM

    2000-12-19

    The Database of Environmental Parameters, Organizations, and Tools (DEPOT) has been developed by the Department of Energy (DOE) as a central warehouse for access to data essential for environmental risk assessment analyses. Initial efforts have concentrated on groundwater and vadose zone transport data and bioaccumulation factors. DEPOT seeks to provide a source of referenced data that, wherever possible, includes the level of uncertainty associated with these parameters. Based on the amount of data available for a particular parameter, uncertainty is expressed as a standard deviation or a distribution function. DEPOT also provides DOE site-specific performance assessment data, pathway-specific transport data, and links to environmental regulations, disposal site waste acceptance criteria, other environmental parameter databases, and environmental risk assessment models.

  5. Data-based mechanistic modeling of dissolved organic carbon load through storms using continuous 15-minute resolution observations within UK upland watersheds

    NASA Astrophysics Data System (ADS)

    Jones, T.; Chappell, N. A.

    2013-12-01

    Few watershed modeling studies have addressed DOC dynamics through storm hydrographs (notable exceptions include Boyer et al., 1997 Hydrol Process; Jutras et al., 2011 Ecol Model; Xu et al., 2012 Water Resour Res). In part this has been a consequence of an incomplete understanding of the biogeochemical processes leading to DOC export to streams (Neff & Asner, 2001, Ecosystems) & an insufficient frequency of DOC monitoring to capture sometimes complex time-varying relationships between DOC & storm hydrographs (Kirchner et al., 2004, Hydrol Process). We present the results of a new & ongoing UK study that integrates two components - 1/ New observations of DOC concentrations (& derived load) continuously monitored at 15 minute intervals through multiple seasons for replicated watersheds; & 2/ A dynamic modeling technique that is able to quantify storage-decay effects, plus hysteretic, nonlinear, lagged & non-stationary relationships between DOC & controlling variables (including rainfall, streamflow, temperature & specific biogeochemical variables e.g., pH, nitrate). DOC concentration is being monitored continuously using the latest generation of UV spectrophotometers (i.e. S::CAN spectro::lysers) with in situ calibrations to laboratory analyzed DOC. The controlling variables are recorded simultaneously at the same stream stations. The watersheds selected for study are among the most intensively studied basins in the UK uplands, namely the Plynlimon & Llyn Brianne experimental basins. All contain areas of organic soils, with three having improved grasslands & three conifer afforested. The dynamic response characteristics (DRCs) that describe detailed DOC behaviour through sequences of storms are simulated using the latest identification routines for continuous time transfer function (CT-TF) models within the Matlab-based CAPTAIN toolbox (some incorporating nonlinear components). To our knowledge this is the first application of CT-TFs to modelling DOC processes

  6. Organ system heterogeneity DB: a database for the visualization of phenotypes at the organ system level

    PubMed Central

    Mannil, Deepthi; Vogt, Ingo; Prinz, Jeanette; Campillos, Monica

    2015-01-01

    Perturbations of mammalian organisms including diseases, drug treatments and gene perturbations in mice affect organ systems differently. Some perturbations impair relatively few organ systems while others lead to highly heterogeneous or systemic effects. Organ System Heterogeneity DB (http://mips.helmholtz-muenchen.de/Organ_System_Heterogeneity/) provides information on the phenotypic effects of 4865 human diseases, 1667 drugs and 5361 genetically modified mouse models on 26 different organ systems. Disease symptoms, drug side effects and mouse phenotypes are mapped to the System Organ Class (SOC) level of the Medical Dictionary of Regulatory Activities (MedDRA). Then, the organ system heterogeneity value, a measurement of the systemic impact of a perturbation, is calculated from the relative frequency of phenotypic features across all SOCs. For perturbations of interest, the database displays the distribution of phenotypic effects across organ systems along with the heterogeneity value and the distance between organ system distributions. In this way, it allows, in an easy and comprehensible fashion, the comparison of the phenotypic organ system distributions of diseases, drugs and their corresponding genetically modified mouse models of associated disease genes and drug targets. The Organ System Heterogeneity DB is thus a platform for the visualization and comparison of organ system level phenotypic effects of drugs, diseases and genes. PMID:25313158

  7. Models And Results Database System.

    2001-03-27

    Version 00 MAR-D 4.16 is a program that is used primarily for Probabilistic Risk Assessment (PRA) data loading. This program defines a common relational database structure that is used by other PRA programs. This structure allows all of the software to access and manipulate data created by other software in the system without performing a lengthy conversion. The MAR-D program also provides the facilities for loading and unloading of PRA data from the relational databasemore » structure used to store the data to an ASCII format for interchange with other PRA software. The primary function of MAR-D is to create a data repository for NUREG-1150 and other permanent data by providing input, conversion, and output capabilities for data used by IRRAS, SARA, SETS and FRANTIC.« less

  8. Web resources for model organism studies.

    PubMed

    Tang, Bixia; Wang, Yanqing; Zhu, Junwei; Zhao, Wenming

    2015-02-01

    An ever-growing number of resources on model organisms have emerged with the continued development of sequencing technologies. In this paper, we review 13 databases of model organisms, most of which are reported by the National Institutes of Health of the United States (NIH; http://www.nih.gov/science/models/). We provide a brief description for each database, as well as detail its data source and types, functions, tools, and availability of access. In addition, we also provide a quality assessment about these databases. Significantly, the organism databases instituted in the early 1990s--such as the Mouse Genome Database (MGD), Saccharomyces Genome Database (SGD), and FlyBase--have developed into what are now comprehensive, core authority resources. Furthermore, all of the databases mentioned here update continually according to user feedback and with advancing technologies. PMID:25707592

  9. Development and mining of a volatile organic compound database.

    PubMed

    Abdullah, Azian Azamimi; Altaf-Ul-Amin, Md; Ono, Naoaki; Sato, Tetsuo; Sugiura, Tadao; Morita, Aki Hirai; Katsuragi, Tetsuo; Muto, Ai; Nishioka, Takaaki; Kanaya, Shigehiko

    2015-01-01

    Volatile organic compounds (VOCs) are small molecules that exhibit high vapor pressure under ambient conditions and have low boiling points. Although VOCs contribute only a small proportion of the total metabolites produced by living organisms, they play an important role in chemical ecology specifically in the biological interactions between organisms and ecosystems. VOCs are also important in the health care field as they are presently used as a biomarker to detect various human diseases. Information on VOCs is scattered in the literature until now; however, there is still no available database describing VOCs and their biological activities. To attain this purpose, we have developed KNApSAcK Metabolite Ecology Database, which contains the information on the relationships between VOCs and their emitting organisms. The KNApSAcK Metabolite Ecology is also linked with the KNApSAcK Core and KNApSAcK Metabolite Activity Database to provide further information on the metabolites and their biological activities. The VOC database can be accessed online. PMID:26495281

  10. Development and mining of a volatile organic compound database.

    PubMed

    Abdullah, Azian Azamimi; Altaf-Ul-Amin, Md; Ono, Naoaki; Sato, Tetsuo; Sugiura, Tadao; Morita, Aki Hirai; Katsuragi, Tetsuo; Muto, Ai; Nishioka, Takaaki; Kanaya, Shigehiko

    2015-01-01

    Volatile organic compounds (VOCs) are small molecules that exhibit high vapor pressure under ambient conditions and have low boiling points. Although VOCs contribute only a small proportion of the total metabolites produced by living organisms, they play an important role in chemical ecology specifically in the biological interactions between organisms and ecosystems. VOCs are also important in the health care field as they are presently used as a biomarker to detect various human diseases. Information on VOCs is scattered in the literature until now; however, there is still no available database describing VOCs and their biological activities. To attain this purpose, we have developed KNApSAcK Metabolite Ecology Database, which contains the information on the relationships between VOCs and their emitting organisms. The KNApSAcK Metabolite Ecology is also linked with the KNApSAcK Core and KNApSAcK Metabolite Activity Database to provide further information on the metabolites and their biological activities. The VOC database can be accessed online.

  11. Development and Mining of a Volatile Organic Compound Database

    PubMed Central

    Abdullah, Azian Azamimi; Altaf-Ul-Amin, Md.; Ono, Naoaki; Sato, Tetsuo; Sugiura, Tadao; Morita, Aki Hirai; Katsuragi, Tetsuo; Muto, Ai; Nishioka, Takaaki; Kanaya, Shigehiko

    2015-01-01

    Volatile organic compounds (VOCs) are small molecules that exhibit high vapor pressure under ambient conditions and have low boiling points. Although VOCs contribute only a small proportion of the total metabolites produced by living organisms, they play an important role in chemical ecology specifically in the biological interactions between organisms and ecosystems. VOCs are also important in the health care field as they are presently used as a biomarker to detect various human diseases. Information on VOCs is scattered in the literature until now; however, there is still no available database describing VOCs and their biological activities. To attain this purpose, we have developed KNApSAcK Metabolite Ecology Database, which contains the information on the relationships between VOCs and their emitting organisms. The KNApSAcK Metabolite Ecology is also linked with the KNApSAcK Core and KNApSAcK Metabolite Activity Database to provide further information on the metabolites and their biological activities. The VOC database can be accessed online. PMID:26495281

  12. The Database for Reaching Experiments and Models

    PubMed Central

    Walker, Ben; Kording, Konrad

    2013-01-01

    Reaching is one of the central experimental paradigms in the field of motor control, and many computational models of reaching have been published. While most of these models try to explain subject data (such as movement kinematics, reaching performance, forces, etc.) from only a single experiment, distinct experiments often share experimental conditions and record similar kinematics. This suggests that reaching models could be applied to (and falsified by) multiple experiments. However, using multiple datasets is difficult because experimental data formats vary widely. Standardizing data formats promises to enable scientists to test model predictions against many experiments and to compare experimental results across labs. Here we report on the development of a new resource available to scientists: a database of reaching called the Database for Reaching Experiments And Models (DREAM). DREAM collects both experimental datasets and models and facilitates their comparison by standardizing formats. The DREAM project promises to be useful for experimentalists who want to understand how their data relates to models, for modelers who want to test their theories, and for educators who want to help students better understand reaching experiments, models, and data analysis. PMID:24244351

  13. Hydroacoustic forcing function modeling using DNS database

    NASA Technical Reports Server (NTRS)

    Zawadzki, I.; Gershfield, J. L.; Na, Y.; Wang, M.

    1996-01-01

    A wall pressure frequency spectrum model (Blake 1971 ) has been evaluated using databases from Direct Numerical Simulations (DNS) of a turbulent boundary layer (Na & Moin 1996). Good agreement is found for moderate to strong adverse pressure gradient flows in the absence of separation. In the separated flow region, the model underpredicts the directly calculated spectra by an order of magnitude. The discrepancy is attributed to the violation of the model assumptions in that part of the flow domain. DNS computed coherence length scales and the normalized wall pressure cross-spectra are compared with experimental data. The DNS results are consistent with experimental observations.

  14. Combining Soil Databases for Topsoil Organic Carbon Mapping in Europe.

    PubMed

    Aksoy, Ece; Yigini, Yusuf; Montanarella, Luca

    2016-01-01

    Accuracy in assessing the distribution of soil organic carbon (SOC) is an important issue because of playing key roles in the functions of both natural ecosystems and agricultural systems. There are several studies in the literature with the aim of finding the best method to assess and map the distribution of SOC content for Europe. Therefore this study aims searching for another aspect of this issue by looking to the performances of using aggregated soil samples coming from different studies and land-uses. The total number of the soil samples in this study was 23,835 and they're collected from the "Land Use/Cover Area frame Statistical Survey" (LUCAS) Project (samples from agricultural soil), BioSoil Project (samples from forest soil), and "Soil Transformations in European Catchments" (SoilTrEC) Project (samples from local soil data coming from six different critical zone observatories (CZOs) in Europe). Moreover, 15 spatial indicators (slope, aspect, elevation, compound topographic index (CTI), CORINE land-cover classification, parent material, texture, world reference base (WRB) soil classification, geological formations, annual average temperature, min-max temperature, total precipitation and average precipitation (for years 1960-1990 and 2000-2010)) were used as auxiliary variables in this prediction. One of the most popular geostatistical techniques, Regression-Kriging (RK), was applied to build the model and assess the distribution of SOC. This study showed that, even though RK method was appropriate for successful SOC mapping, using combined databases was not helpful to increase the statistical significance of the method results for assessing the SOC distribution. According to our results; SOC variation was mainly affected by elevation, slope, CTI, average temperature, average and total precipitation, texture, WRB and CORINE variables for Europe scale in our model. Moreover, the highest average SOC contents were found in the wetland areas; agricultural

  15. Combining Soil Databases for Topsoil Organic Carbon Mapping in Europe.

    PubMed

    Aksoy, Ece; Yigini, Yusuf; Montanarella, Luca

    2016-01-01

    Accuracy in assessing the distribution of soil organic carbon (SOC) is an important issue because of playing key roles in the functions of both natural ecosystems and agricultural systems. There are several studies in the literature with the aim of finding the best method to assess and map the distribution of SOC content for Europe. Therefore this study aims searching for another aspect of this issue by looking to the performances of using aggregated soil samples coming from different studies and land-uses. The total number of the soil samples in this study was 23,835 and they're collected from the "Land Use/Cover Area frame Statistical Survey" (LUCAS) Project (samples from agricultural soil), BioSoil Project (samples from forest soil), and "Soil Transformations in European Catchments" (SoilTrEC) Project (samples from local soil data coming from six different critical zone observatories (CZOs) in Europe). Moreover, 15 spatial indicators (slope, aspect, elevation, compound topographic index (CTI), CORINE land-cover classification, parent material, texture, world reference base (WRB) soil classification, geological formations, annual average temperature, min-max temperature, total precipitation and average precipitation (for years 1960-1990 and 2000-2010)) were used as auxiliary variables in this prediction. One of the most popular geostatistical techniques, Regression-Kriging (RK), was applied to build the model and assess the distribution of SOC. This study showed that, even though RK method was appropriate for successful SOC mapping, using combined databases was not helpful to increase the statistical significance of the method results for assessing the SOC distribution. According to our results; SOC variation was mainly affected by elevation, slope, CTI, average temperature, average and total precipitation, texture, WRB and CORINE variables for Europe scale in our model. Moreover, the highest average SOC contents were found in the wetland areas; agricultural

  16. Combining Soil Databases for Topsoil Organic Carbon Mapping in Europe

    PubMed Central

    Aksoy, Ece

    2016-01-01

    Accuracy in assessing the distribution of soil organic carbon (SOC) is an important issue because of playing key roles in the functions of both natural ecosystems and agricultural systems. There are several studies in the literature with the aim of finding the best method to assess and map the distribution of SOC content for Europe. Therefore this study aims searching for another aspect of this issue by looking to the performances of using aggregated soil samples coming from different studies and land-uses. The total number of the soil samples in this study was 23,835 and they’re collected from the “Land Use/Cover Area frame Statistical Survey” (LUCAS) Project (samples from agricultural soil), BioSoil Project (samples from forest soil), and “Soil Transformations in European Catchments” (SoilTrEC) Project (samples from local soil data coming from six different critical zone observatories (CZOs) in Europe). Moreover, 15 spatial indicators (slope, aspect, elevation, compound topographic index (CTI), CORINE land-cover classification, parent material, texture, world reference base (WRB) soil classification, geological formations, annual average temperature, min-max temperature, total precipitation and average precipitation (for years 1960–1990 and 2000–2010)) were used as auxiliary variables in this prediction. One of the most popular geostatistical techniques, Regression-Kriging (RK), was applied to build the model and assess the distribution of SOC. This study showed that, even though RK method was appropriate for successful SOC mapping, using combined databases was not helpful to increase the statistical significance of the method results for assessing the SOC distribution. According to our results; SOC variation was mainly affected by elevation, slope, CTI, average temperature, average and total precipitation, texture, WRB and CORINE variables for Europe scale in our model. Moreover, the highest average SOC contents were found in the wetland areas

  17. Spatial Database Modeling for Indoor Navigation Systems

    NASA Astrophysics Data System (ADS)

    Gotlib, Dariusz; Gnat, Miłosz

    2013-12-01

    For many years, cartographers are involved in designing GIS and navigation systems. Most GIS applications use the outdoor data. Increasingly, similar applications are used inside buildings. Therefore it is important to find the proper model of indoor spatial database. The development of indoor navigation systems should utilize advanced teleinformation, geoinformatics, geodetic and cartographical knowledge. The authors present the fundamental requirements for the indoor data model for navigation purposes. Presenting some of the solutions adopted in the world they emphasize that navigation applications require specific data to present the navigation routes in the right way. There is presented original solution for indoor data model created by authors on the basis of BISDM model. Its purpose is to expand the opportunities for use in indoor navigation.

  18. Assessment of the SFC database for analysis and modeling

    NASA Technical Reports Server (NTRS)

    Centeno, Martha A.

    1994-01-01

    SFC is one of the four clusters that make up the Integrated Work Control System (IWCS), which will integrate the shuttle processing databases at Kennedy Space Center (KSC). The IWCS framework will enable communication among the four clusters and add new data collection protocols. The Shop Floor Control (SFC) module has been operational for two and a half years; however, at this stage, automatic links to the other 3 modules have not been implemented yet, except for a partial link to IOS (CASPR). SFC revolves around a DB/2 database with PFORMS acting as the database management system (DBMS). PFORMS is an off-the-shelf DB/2 application that provides a set of data entry screens and query forms. The main dynamic entity in the SFC and IOS database is a task; thus, the physical storage location and update privileges are driven by the status of the WAD. As we explored the SFC values, we realized that there was much to do before actually engaging in continuous analysis of the SFC data. Half way into this effort, it was realized that full scale analysis would have to be a future third phase of this effort. So, we concentrated on getting to know the contents of the database, and in establishing an initial set of tools to start the continuous analysis process. Specifically, we set out to: (1) provide specific procedures for statistical models, so as to enhance the TP-OAO office analysis and modeling capabilities; (2) design a data exchange interface; (3) prototype the interface to provide inputs to SCRAM; and (4) design a modeling database. These objectives were set with the expectation that, if met, they would provide former TP-OAO engineers with tools that would help them demonstrate the importance of process-based analyses. The latter, in return, will help them obtain the cooperation of various organizations in charting out their individual processes.

  19. Asteroid models from the Lowell photometric database

    NASA Astrophysics Data System (ADS)

    Ďurech, J.; Hanuš, J.; Oszkiewicz, D.; Vančo, R.

    2016-03-01

    Context. Information about shapes and spin states of individual asteroids is important for the study of the whole asteroid population. For asteroids from the main belt, most of the shape models available now have been reconstructed from disk-integrated photometry by the lightcurve inversion method. Aims: We want to significantly enlarge the current sample (~350) of available asteroid models. Methods: We use the lightcurve inversion method to derive new shape models and spin states of asteroids from the sparse-in-time photometry compiled in the Lowell Photometric Database. To speed up the time-consuming process of scanning the period parameter space through the use of convex shape models, we use the distributed computing project Asteroids@home, running on the Berkeley Open Infrastructure for Network Computing (BOINC) platform. This way, the period-search interval is divided into hundreds of smaller intervals. These intervals are scanned separately by different volunteers and then joined together. We also use an alternative, faster, approach when searching the best-fit period by using a model of triaxial ellipsoid. By this, we can independently confirm periods found with convex models and also find rotation periods for some of those asteroids for which the convex-model approach gives too many solutions. Results: From the analysis of Lowell photometric data of the first 100 000 numbered asteroids, we derived 328 new models. This almost doubles the number of available models. We tested the reliability of our results by comparing models that were derived from purely Lowell data with those based on dense lightcurves, and we found that the rate of false-positive solutions is very low. We also present updated plots of the distribution of spin obliquities and pole ecliptic longitudes that confirm previous findings about a non-uniform distribution of spin axes. However, the models reconstructed from noisy sparse data are heavily biased towards more elongated bodies with high

  20. Integrated Space Asset Management Database and Modeling

    NASA Technical Reports Server (NTRS)

    MacLeod, Todd; Gagliano, Larry; Percy, Thomas; Mason, Shane

    2015-01-01

    Effective Space Asset Management is one key to addressing the ever-growing issue of space congestion. It is imperative that agencies around the world have access to data regarding the numerous active assets and pieces of space junk currently tracked in orbit around the Earth. At the center of this issues is the effective management of data of many types related to orbiting objects. As the population of tracked objects grows, so too should the data management structure used to catalog technical specifications, orbital information, and metadata related to those populations. Marshall Space Flight Center's Space Asset Management Database (SAM-D) was implemented in order to effectively catalog a broad set of data related to known objects in space by ingesting information from a variety of database and processing that data into useful technical information. Using the universal NORAD number as a unique identifier, the SAM-D processes two-line element data into orbital characteristics and cross-references this technical data with metadata related to functional status, country of ownership, and application category. The SAM-D began as an Excel spreadsheet and was later upgraded to an Access database. While SAM-D performs its task very well, it is limited by its current platform and is not available outside of the local user base. Further, while modeling and simulation can be powerful tools to exploit the information contained in SAM-D, the current system does not allow proper integration options for combining the data with both legacy and new M&S tools. This paper provides a summary of SAM-D development efforts to date and outlines a proposed data management infrastructure that extends SAM-D to support the larger data sets to be generated. A service-oriented architecture model using an information sharing platform named SIMON will allow it to easily expand to incorporate new capabilities, including advanced analytics, M&S tools, fusion techniques and user interface for

  1. A computational framework for a database of terrestrial biosphere models

    NASA Astrophysics Data System (ADS)

    Metzler, Holger; Müller, Markus; Ceballos-Núñez, Verónika; Sierra, Carlos A.

    2016-04-01

    Most terrestrial biosphere models consist of a set of coupled ordinary first order differential equations. Each equation represents a pool containing carbon with a certain turnover rate. Although such models share some basic mathematical structures, they can have very different properties such as number of pools, cycling rates, and internal fluxes. We present a computational framework that helps analyze the structure and behavior of terrestrial biosphere models using as an example the process of soil organic matter decomposition. The same framework can also be used for other sub-processes such as carbon fixation or allocation. First, the models have to be fed into a database consisting of simple text files with a common structure. Then they are read in using Python and transformed into an internal 'Model Class' that can be used to automatically create an overview stating the model's structure, state variables, internal and external fluxes. SymPy, a Python library for symbolic mathematics, helps to also calculate the Jacobian matrix at possibly given steady states and the eigenvalues of this matrix. If complete parameter sets are available, the model can also be run using R to simulate its behavior under certain conditions and to support a deeper stability analysis. In this case, the framework is also able to provide phase-plane plots if appropriate. Furthermore, an overview of all the models in the database can be given to help identify their similarities and differences.

  2. Techniques to Access Databases and Integrate Data for Hydrologic Modeling

    SciTech Connect

    Whelan, Gene; Tenney, Nathan D.; Pelton, Mitchell A.; Coleman, Andre M.; Ward, Duane L.; Droppo, James G.; Meyer, Philip D.; Dorow, Kevin E.; Taira, Randal Y.

    2009-06-17

    This document addresses techniques to access and integrate data for defining site-specific conditions and behaviors associated with ground-water and surface-water radionuclide transport applicable to U.S. Nuclear Regulatory Commission reviews. Environmental models typically require input data from multiple internal and external sources that may include, but are not limited to, stream and rainfall gage data, meteorological data, hydrogeological data, habitat data, and biological data. These data may be retrieved from a variety of organizations (e.g., federal, state, and regional) and source types (e.g., HTTP, FTP, and databases). Available data sources relevant to hydrologic analyses for reactor licensing are identified and reviewed. The data sources described can be useful to define model inputs and parameters, including site features (e.g., watershed boundaries, stream locations, reservoirs, site topography), site properties (e.g., surface conditions, subsurface hydraulic properties, water quality), and site boundary conditions, input forcings, and extreme events (e.g., stream discharge, lake levels, precipitation, recharge, flood and drought characteristics). Available software tools for accessing established databases, retrieving the data, and integrating it with models were identified and reviewed. The emphasis in this review was on existing software products with minimal required modifications to enable their use with the FRAMES modeling framework. The ability of four of these tools to access and retrieve the identified data sources was reviewed. These four software tools were the Hydrologic Data Acquisition and Processing System (HDAPS), Integrated Water Resources Modeling System (IWRMS) External Data Harvester, Data for Environmental Modeling Environmental Data Download Tool (D4EM EDDT), and the FRAMES Internet Database Tools. The IWRMS External Data Harvester and the D4EM EDDT were identified as the most promising tools based on their ability to access and

  3. Fish Karyome version 2.1: a chromosome database of fishes and other aquatic organisms.

    PubMed

    Nagpure, Naresh Sahebrao; Pathak, Ajey Kumar; Pati, Rameshwar; Rashid, Iliyas; Sharma, Jyoti; Singh, Shri Prakash; Singh, Mahender; Sarkar, Uttam Kumar; Kushwaha, Basdeo; Kumar, Ravindra; Murali, S

    2016-01-01

    A voluminous information is available on karyological studies of fishes; however, limited efforts were made for compilation and curation of the available karyological data in a digital form. 'Fish Karyome' database was the preliminary attempt to compile and digitize the available karyological information on finfishes belonging to the Indian subcontinent. But the database had limitations since it covered data only on Indian finfishes with limited search options. Perceiving the feedbacks from the users and its utility in fish cytogenetic studies, the Fish Karyome database was upgraded by applying Linux, Apache, MySQL and PHP (pre hypertext processor) (LAMP) technologies. In the present version, the scope of the system was increased by compiling and curating the available chromosomal information over the globe on fishes and other aquatic organisms, such as echinoderms, molluscs and arthropods, especially of aquaculture importance. Thus, Fish Karyome version 2.1 presently covers 866 chromosomal records for 726 species supported with 253 published articles and the information is being updated regularly. The database provides information on chromosome number and morphology, sex chromosomes, chromosome banding, molecular cytogenetic markers, etc. supported by fish and karyotype images through interactive tools. It also enables the users to browse and view chromosomal information based on habitat, family, conservation status and chromosome number. The system also displays chromosome number in model organisms, protocol for chromosome preparation and allied techniques and glossary of cytogenetic terms. A data submission facility has also been provided through data submission panel. The database can serve as a unique and useful resource for cytogenetic characterization, sex determination, chromosomal mapping, cytotaxonomy, karyo-evolution and systematics of fishes. Database URL: http://mail.nbfgr.res.in/Fish_Karyome.

  4. Fish Karyome version 2.1: a chromosome database of fishes and other aquatic organisms

    PubMed Central

    Nagpure, Naresh Sahebrao; Pathak, Ajey Kumar; Pati, Rameshwar; Rashid, Iliyas; Sharma, Jyoti; Singh, Shri Prakash; Singh, Mahender; Sarkar, Uttam Kumar; Kushwaha, Basdeo; Kumar, Ravindra; Murali, S.

    2016-01-01

    A voluminous information is available on karyological studies of fishes; however, limited efforts were made for compilation and curation of the available karyological data in a digital form. ‘Fish Karyome’ database was the preliminary attempt to compile and digitize the available karyological information on finfishes belonging to the Indian subcontinent. But the database had limitations since it covered data only on Indian finfishes with limited search options. Perceiving the feedbacks from the users and its utility in fish cytogenetic studies, the Fish Karyome database was upgraded by applying Linux, Apache, MySQL and PHP (pre hypertext processor) (LAMP) technologies. In the present version, the scope of the system was increased by compiling and curating the available chromosomal information over the globe on fishes and other aquatic organisms, such as echinoderms, molluscs and arthropods, especially of aquaculture importance. Thus, Fish Karyome version 2.1 presently covers 866 chromosomal records for 726 species supported with 253 published articles and the information is being updated regularly. The database provides information on chromosome number and morphology, sex chromosomes, chromosome banding, molecular cytogenetic markers, etc. supported by fish and karyotype images through interactive tools. It also enables the users to browse and view chromosomal information based on habitat, family, conservation status and chromosome number. The system also displays chromosome number in model organisms, protocol for chromosome preparation and allied techniques and glossary of cytogenetic terms. A data submission facility has also been provided through data submission panel. The database can serve as a unique and useful resource for cytogenetic characterization, sex determination, chromosomal mapping, cytotaxonomy, karyo-evolution and systematics of fishes. Database URL: http://mail.nbfgr.res.in/Fish_Karyome PMID:26980518

  5. Fish Karyome version 2.1: a chromosome database of fishes and other aquatic organisms.

    PubMed

    Nagpure, Naresh Sahebrao; Pathak, Ajey Kumar; Pati, Rameshwar; Rashid, Iliyas; Sharma, Jyoti; Singh, Shri Prakash; Singh, Mahender; Sarkar, Uttam Kumar; Kushwaha, Basdeo; Kumar, Ravindra; Murali, S

    2016-01-01

    A voluminous information is available on karyological studies of fishes; however, limited efforts were made for compilation and curation of the available karyological data in a digital form. 'Fish Karyome' database was the preliminary attempt to compile and digitize the available karyological information on finfishes belonging to the Indian subcontinent. But the database had limitations since it covered data only on Indian finfishes with limited search options. Perceiving the feedbacks from the users and its utility in fish cytogenetic studies, the Fish Karyome database was upgraded by applying Linux, Apache, MySQL and PHP (pre hypertext processor) (LAMP) technologies. In the present version, the scope of the system was increased by compiling and curating the available chromosomal information over the globe on fishes and other aquatic organisms, such as echinoderms, molluscs and arthropods, especially of aquaculture importance. Thus, Fish Karyome version 2.1 presently covers 866 chromosomal records for 726 species supported with 253 published articles and the information is being updated regularly. The database provides information on chromosome number and morphology, sex chromosomes, chromosome banding, molecular cytogenetic markers, etc. supported by fish and karyotype images through interactive tools. It also enables the users to browse and view chromosomal information based on habitat, family, conservation status and chromosome number. The system also displays chromosome number in model organisms, protocol for chromosome preparation and allied techniques and glossary of cytogenetic terms. A data submission facility has also been provided through data submission panel. The database can serve as a unique and useful resource for cytogenetic characterization, sex determination, chromosomal mapping, cytotaxonomy, karyo-evolution and systematics of fishes. Database URL: http://mail.nbfgr.res.in/Fish_Karyome. PMID:26980518

  6. Sequence modelling and an extensible data model for genomic database

    SciTech Connect

    Li, Peter Wei-Der Lawrence Berkeley Lab., CA )

    1992-01-01

    The Human Genome Project (HGP) plans to sequence the human genome by the beginning of the next century. It will generate DNA sequences of more than 10 billion bases and complex marker sequences (maps) of more than 100 million markers. All of these information will be stored in database management systems (DBMSs). However, existing data models do not have the abstraction mechanism for modelling sequences and existing DBMS's do not have operations for complex sequences. This work addresses the problem of sequence modelling in the context of the HGP and the more general problem of an extensible object data model that can incorporate the sequence model as well as existing and future data constructs and operators. First, we proposed a general sequence model that is application and implementation independent. This model is used to capture the sequence information found in the HGP at the conceptual level. In addition, abstract and biological sequence operators are defined for manipulating the modelled sequences. Second, we combined many features of semantic and object oriented data models into an extensible framework, which we called the Extensible Object Model'', to address the need of a modelling framework for incorporating the sequence data model with other types of data constructs and operators. This framework is based on the conceptual separation between constructors and constraints. We then used this modelling framework to integrate the constructs for the conceptual sequence model. The Extensible Object Model is also defined with a graphical representation, which is useful as a tool for database designers. Finally, we defined a query language to support this model and implement the query processor to demonstrate the feasibility of the extensible framework and the usefulness of the conceptual sequence model.

  7. Sequence modelling and an extensible data model for genomic database

    SciTech Connect

    Li, Peter Wei-Der |

    1992-01-01

    The Human Genome Project (HGP) plans to sequence the human genome by the beginning of the next century. It will generate DNA sequences of more than 10 billion bases and complex marker sequences (maps) of more than 100 million markers. All of these information will be stored in database management systems (DBMSs). However, existing data models do not have the abstraction mechanism for modelling sequences and existing DBMS`s do not have operations for complex sequences. This work addresses the problem of sequence modelling in the context of the HGP and the more general problem of an extensible object data model that can incorporate the sequence model as well as existing and future data constructs and operators. First, we proposed a general sequence model that is application and implementation independent. This model is used to capture the sequence information found in the HGP at the conceptual level. In addition, abstract and biological sequence operators are defined for manipulating the modelled sequences. Second, we combined many features of semantic and object oriented data models into an extensible framework, which we called the ``Extensible Object Model``, to address the need of a modelling framework for incorporating the sequence data model with other types of data constructs and operators. This framework is based on the conceptual separation between constructors and constraints. We then used this modelling framework to integrate the constructs for the conceptual sequence model. The Extensible Object Model is also defined with a graphical representation, which is useful as a tool for database designers. Finally, we defined a query language to support this model and implement the query processor to demonstrate the feasibility of the extensible framework and the usefulness of the conceptual sequence model.

  8. Organ growth functions in maturing male Sprague-Dawley rats based on a collective database.

    PubMed

    Mirfazaelian, Ahmad; Fisher, Jeffrey W

    2007-06-01

    Ten different organ weights (liver, spleen, kidneys, heart, lungs, brain, adrenals, testes, epididymes, and seminal vesicles) of male Sprague-Dawley (S-D) rats of different ages (1-280 d) were extracted based on a thorough literature survey database. A generalized Michaelis-Menten (GMM) model, used to fit organ weights versus age in a previous study (Schoeffner et al., 1999) based on a limited data, was used to find the best fit model for the present expanded data compilation. The GMM model has the functional form: Wt = (Wt(o).K(gamma) + Wt(max).Age(gamma))/(K(gamma) + Age(gamma)) where Wt is organ/tissue weight at a specified age, Wt(o) and Wt(max) are weight at birth and maximal growth, respectively, and K and gamma are constants. Organ weights were significantly correlated with their respective ages for all organs and tissues. GMM-derived organ growth and percent body weight (%BW) fractions of different tissues were plotted against animal age and compared with experimental values as well as previously published models. The GMM-based organ growth and %BW fraction profiles were in general agreement with our empirical data as well as with previous studies. The present model was compared with the GMM model developed previously for six organs--liver, spleen, kidneys, heart, lungs, and brain--based on a limited data, and no significant difference was noticed between the two sets of predictions. It was concluded that the GMM models presented herein for different male S-D rats organs (liver, spleen, kidneys, heart, lungs, brain, adrenals, testes, epididymes, and seminal vesicles) are capable of predicting organ weights and %BW ratios accurately at different ages. PMID:17497417

  9. Solid Waste Projection Model: Database (Version 1. 3)

    SciTech Connect

    Blackburn, C.L.

    1991-11-01

    The Solid Waste Projection Model (SWPM) system is an analytical tool developed by Pacific Northwest Laboratory (PNL) for Westinghouse Hanford Company (WHC). The SWPM system provides a modeling and analysis environment that supports decisions in the process of evaluating various solid waste management alternatives. This document, one of a series describing the SWPM system, contains detailed information regarding the software and data structures utilized in developing the SWPM Version 1.3 Database. This document is intended for use by experienced database specialists and supports database maintenance, utility development, and database enhancement.

  10. Software Engineering Laboratory (SEL) database organization and user's guide, revision 2

    NASA Technical Reports Server (NTRS)

    Morusiewicz, Linda; Bristow, John

    1992-01-01

    The organization of the Software Engineering Laboratory (SEL) database is presented. Included are definitions and detailed descriptions of the database tables and views, the SEL data, and system support data. The mapping from the SEL and system support data to the base table is described. In addition, techniques for accessing the database through the Database Access Manager for the SEL (DAMSEL) system and via the ORACLE structured query language (SQL) are discussed.

  11. Integrated Space Asset Management Database and Modeling

    NASA Astrophysics Data System (ADS)

    Gagliano, L.; MacLeod, T.; Mason, S.; Percy, T.; Prescott, J.

    The Space Asset Management Database (SAM-D) was implemented in order to effectively track known objects in space by ingesting information from a variety of databases and performing calculations to determine the expected position of the object at a specified time. While SAM-D performs this task very well, it is limited by technology and is not available outside of the local user base. Modeling and simulation can be powerful tools to exploit the information contained in SAM-D. However, the current system does not allow proper integration options for combining the data with both legacy and new M&S tools. A more capable data management infrastructure would extend SAM-D to support the larger data sets to be generated by the COI. A service-oriented architecture model will allow it to easily expand to incorporate new capabilities, including advanced analytics, M&S tools, fusion techniques and user interface for visualizations. Based on a web-centric approach, the entire COI will be able to access the data and related analytics. In addition, tight control of information sharing policy will increase confidence in the system, which would encourage industry partners to provide commercial data. SIMON is a Government off the Shelf information sharing platform in use throughout DoD and DHS information sharing and situation awareness communities. SIMON providing fine grained control to data owners allowing them to determine exactly how and when their data is shared. SIMON supports a micro-service approach to system development, meaning M&S and analytic services can be easily built or adapted. It is uniquely positioned to fill this need as an information-sharing platform with a proven track record of successful situational awareness system deployments. Combined with the integration of new and legacy M&S tools, a SIMON-based architecture will provide a robust SA environment for the NASA SA COI that can be extended and expanded indefinitely. First Results of Coherent Uplink from a

  12. Integrated Space Asset Management Database and Modeling

    NASA Astrophysics Data System (ADS)

    Gagliano, L.; MacLeod, T.; Mason, S.; Percy, T.; Prescott, J.

    The Space Asset Management Database (SAM-D) was implemented in order to effectively track known objects in space by ingesting information from a variety of databases and performing calculations to determine the expected position of the object at a specified time. While SAM-D performs this task very well, it is limited by technology and is not available outside of the local user base. Modeling and simulation can be powerful tools to exploit the information contained in SAM-D. However, the current system does not allow proper integration options for combining the data with both legacy and new M&S tools. A more capable data management infrastructure would extend SAM-D to support the larger data sets to be generated by the COI. A service-oriented architecture model will allow it to easily expand to incorporate new capabilities, including advanced analytics, M&S tools, fusion techniques and user interface for visualizations. Based on a web-centric approach, the entire COI will be able to access the data and related analytics. In addition, tight control of information sharing policy will increase confidence in the system, which would encourage industry partners to provide commercial data. SIMON is a Government off the Shelf information sharing platform in use throughout DoD and DHS information sharing and situation awareness communities. SIMON providing fine grained control to data owners allowing them to determine exactly how and when their data is shared. SIMON supports a micro-service approach to system development, meaning M&S and analytic services can be easily built or adapted. It is uniquely positioned to fill this need as an information-sharing platform with a proven track record of successful situational awareness system deployments. Combined with the integration of new and legacy M&S tools, a SIMON-based architecture will provide a robust SA environment for the NASA SA COI that can be extended and expanded indefinitely. First Results of Coherent Uplink from a

  13. Nonparametric Bayesian Modeling for Automated Database Schema Matching

    SciTech Connect

    Ferragut, Erik M; Laska, Jason A

    2015-01-01

    The problem of merging databases arises in many government and commercial applications. Schema matching, a common first step, identifies equivalent fields between databases. We introduce a schema matching framework that builds nonparametric Bayesian models for each field and compares them by computing the probability that a single model could have generated both fields. Our experiments show that our method is more accurate and faster than the existing instance-based matching algorithms in part because of the use of nonparametric Bayesian models.

  14. Cyclebase 3.0: a multi-organism database on cell-cycle regulation and phenotypes

    PubMed Central

    Santos, Alberto; Wernersson, Rasmus; Jensen, Lars Juhl

    2015-01-01

    The eukaryotic cell division cycle is a highly regulated process that consists of a complex series of events and involves thousands of proteins. Researchers have studied the regulation of the cell cycle in several organisms, employing a wide range of high-throughput technologies, such as microarray-based mRNA expression profiling and quantitative proteomics. Due to its complexity, the cell cycle can also fail or otherwise change in many different ways if important genes are knocked out, which has been studied in several microscopy-based knockdown screens. The data from these many large-scale efforts are not easily accessed, analyzed and combined due to their inherent heterogeneity. To address this, we have created Cyclebase—available at http://www.cyclebase.org—an online database that allows users to easily visualize and download results from genome-wide cell-cycle-related experiments. In Cyclebase version 3.0, we have updated the content of the database to reflect changes to genome annotation, added new mRNA and protein expression data, and integrated cell-cycle phenotype information from high-content screens and model-organism databases. The new version of Cyclebase also features a new web interface, designed around an overview figure that summarizes all the cell-cycle-related data for a gene. PMID:25378319

  15. Cyclebase 3.0: a multi-organism database on cell-cycle regulation and phenotypes.

    PubMed

    Santos, Alberto; Wernersson, Rasmus; Jensen, Lars Juhl

    2015-01-01

    The eukaryotic cell division cycle is a highly regulated process that consists of a complex series of events and involves thousands of proteins. Researchers have studied the regulation of the cell cycle in several organisms, employing a wide range of high-throughput technologies, such as microarray-based mRNA expression profiling and quantitative proteomics. Due to its complexity, the cell cycle can also fail or otherwise change in many different ways if important genes are knocked out, which has been studied in several microscopy-based knockdown screens. The data from these many large-scale efforts are not easily accessed, analyzed and combined due to their inherent heterogeneity. To address this, we have created Cyclebase--available at http://www.cyclebase.org--an online database that allows users to easily visualize and download results from genome-wide cell-cycle-related experiments. In Cyclebase version 3.0, we have updated the content of the database to reflect changes to genome annotation, added new mRNA and protein expression data, and integrated cell-cycle phenotype information from high-content screens and model-organism databases. The new version of Cyclebase also features a new web interface, designed around an overview figure that summarizes all the cell-cycle-related data for a gene.

  16. Evaluation of an SQL model of the HELP patient database.

    PubMed

    Huff, S M; Berthelsen, C L; Pryor, T A; Dudley, A S

    1991-01-01

    We tested a new model of the HELP patient database that makes use of relational tables to store patient data and provides access to data using SQL (Structured Query Language). The SQL database required more storage space and had many more physical records than the HELP database, but it was faster and more efficient in storing data than the standard HELP utilities. The HELP utilities used disk space more efficiently and were faster than the SQL tools when retrieving data for typical clinical reports. However, the SQL model provides networking capabilities, general report writing tools, detailed user documentation, and an ability for creating secondary indexes that offset its poorer performance.

  17. Outline of a Model for Lexical Databases.

    ERIC Educational Resources Information Center

    Ide, Nancy; And Others

    1993-01-01

    Reports on a project showing that relational data models, including unnormalized models that allow the nesting of relations, cannot fully capture the structural properties of lexical information. A feature-based model that allows for a full representation of some nesting and defines a factoring mechanism is described and demonstrated. (38…

  18. Effects of distributed database modeling on evaluation of transaction rollbacks

    NASA Technical Reports Server (NTRS)

    Mukkamala, Ravi

    1991-01-01

    Data distribution, degree of data replication, and transaction access patterns are key factors in determining the performance of distributed database systems. In order to simplify the evaluation of performance measures, database designers and researchers tend to make simplistic assumptions about the system. The effect is studied of modeling assumptions on the evaluation of one such measure, the number of transaction rollbacks, in a partitioned distributed database system. Six probabilistic models and expressions are developed for the numbers of rollbacks under each of these models. Essentially, the models differ in terms of the available system information. The analytical results so obtained are compared to results from simulation. From here, it is concluded that most of the probabilistic models yield overly conservative estimates of the number of rollbacks. The effect of transaction commutativity on system throughout is also grossly undermined when such models are employed.

  19. Human Thermal Model Evaluation Using the JSC Human Thermal Database

    NASA Technical Reports Server (NTRS)

    Bue, Grant; Makinen, Janice; Cognata, Thomas

    2012-01-01

    Human thermal modeling has considerable long term utility to human space flight. Such models provide a tool to predict crew survivability in support of vehicle design and to evaluate crew response in untested space environments. It is to the benefit of any such model not only to collect relevant experimental data to correlate it against, but also to maintain an experimental standard or benchmark for future development in a readily and rapidly searchable and software accessible format. The Human thermal database project is intended to do just so; to collect relevant data from literature and experimentation and to store the data in a database structure for immediate and future use as a benchmark to judge human thermal models against, in identifying model strengths and weakness, to support model development and improve correlation, and to statistically quantify a model s predictive quality. The human thermal database developed at the Johnson Space Center (JSC) is intended to evaluate a set of widely used human thermal models. This set includes the Wissler human thermal model, a model that has been widely used to predict the human thermoregulatory response to a variety of cold and hot environments. These models are statistically compared to the current database, which contains experiments of human subjects primarily in air from a literature survey ranging between 1953 and 2004 and from a suited experiment recently performed by the authors, for a quantitative study of relative strength and predictive quality of the models.

  20. Imprecision and Uncertainty in the UFO Database Model.

    ERIC Educational Resources Information Center

    Van Gyseghem, Nancy; De Caluwe, Rita

    1998-01-01

    Discusses how imprecision and uncertainty are dealt with in the UFO (Uncertainty and Fuzziness in an Object-oriented) database model. Such information is expressed by means of possibility distributions, and modeled by means of the proposed concept of "role objects." The role objects model uncertain, tentative information about objects, and thus…

  1. SIFT vehicle recognition with semi-synthetic model database

    NASA Astrophysics Data System (ADS)

    Price, Rebecca L.; Rovito, Todd V.

    2012-06-01

    Object recognition is an important problem that has many applications that are of interest to the United States Air Force (USAF). Recently the USAF released its update to Technology Horizons, a report that is designed to guide the science and technology direction of the Air Force. Technology Horizons specifically calls out for the need to use autonomous systems in essentially all aspects of Air Force operations [1]. Object recognition is a key enabler to autonomous exploitation of intelligence, surveillance, and reconnaissance (ISR) data which might make the automatic searching of millions of hours of video practical. In particular this paper focuses on vehicle recognition with Lowe's Scale-invariant feature transform (SIFT) using a model database that was generated with semi-synthetic data. To create the model database we used a desktop laser scanner to create a high resolution 3D facet model. Then the 3D facet model was imported into LuxRender, a physics accurate ray tracing tool, and several views were rendered to create a model database. SIFT was selected because the algorithm is invariant to scale, noise, and illumination making it possible to create a model database of only a hundred original viewing locations which keeps the size of the model database reasonable.

  2. Performance modeling for large database systems

    NASA Astrophysics Data System (ADS)

    Schaar, Stephen; Hum, Frank; Romano, Joe

    1997-02-01

    One of the unique approaches Science Applications International Corporation took to meet performance requirements was to start the modeling effort during the proposal phase of the Interstate Identification Index/Federal Bureau of Investigations (III/FBI) project. The III/FBI Performance Model uses analytical modeling techniques to represent the III/FBI system. Inputs to the model include workloads for each transaction type, record size for each record type, number of records for each file, hardware envelope characteristics, engineering margins and estimates for software instructions, memory, and I/O for each transaction type. The model uses queuing theory to calculate the average transaction queue length. The model calculates a response time and the resources needed for each transaction type. Outputs of the model include the total resources needed for the system, a hardware configuration, and projected inherent and operational availability. The III/FBI Performance Model is used to evaluate what-if scenarios and allows a rapid response to engineering change proposals and technical enhancements.

  3. Semantic models in medical record data-bases.

    PubMed

    Cerutti, S

    1980-01-01

    A great effort has been recently made in the area of data-base design in a number of application fields (banking, insurance, travel, etc.). Yet, it is the current experience of computer scientists in the medical field that medical record information-processing requires less rigid and more complete definition of data-base specifications for a much more heterogeneous set of data, for different users who have different aims. Hence, it is important to state that the data-base in the medical field ought to be a model of the environment for which it was created, rather than just a collection of data. New more powerful and more flexible data-base models are being now designed, particularly in the USA, where the current trend in medicine is to implement, in the same structure, the connection among more different and specific users and the data-base (for administrative aims, medical care control, treatments, statistical and epidemiological results, etc.). In such a way the single users are able to talk with the data-base without interfering with one another. The present paper outlines that this multi-purpose flexibility can be achieved by improving mainly the capabilities of the data-base model. This concept allows the creation of procedures of semantic integrity control which will certainly have in the future a dramatic impact on important management features, starting from data-quality checking and non-physiological state detections, as far as more medical-oriented procedures like drug interactions, record surveillance and medical care review. That is especially true when a large amount of data are to be processed and the classical hierarchical and network data models are no longer sufficient for developing satisfactory and reliable automatic procedures. In this regard, particular emphasis will be dedicated to the relational model and, at the highest level, to the same semantic data model.

  4. A Database Model for Medical Consultation.

    ERIC Educational Resources Information Center

    Anvari, Morteza

    1991-01-01

    Describes a relational data model that can be used for knowledge representation and manipulation in rule-based medical consultation systems. Fuzzy queries or attribute values and fuzzy set theory are discussed, functional dependencies are described, and an example is presented of a system for diagnosing causes of eye inflammation. (15 references)…

  5. Human Thermal Model Evaluation Using the JSC Human Thermal Database

    NASA Technical Reports Server (NTRS)

    Cognata, T.; Bue, G.; Makinen, J.

    2011-01-01

    The human thermal database developed at the Johnson Space Center (JSC) is used to evaluate a set of widely used human thermal models. This database will facilitate a more accurate evaluation of human thermoregulatory response using in a variety of situations, including those situations that might otherwise prove too dangerous for actual testing--such as extreme hot or cold splashdown conditions. This set includes the Wissler human thermal model, a model that has been widely used to predict the human thermoregulatory response to a variety of cold and hot environments. These models are statistically compared to the current database, which contains experiments of human subjects primarily in air from a literature survey ranging between 1953 and 2004 and from a suited experiment recently performed by the authors, for a quantitative study of relative strength and predictive quality of the models. Human thermal modeling has considerable long term utility to human space flight. Such models provide a tool to predict crew survivability in support of vehicle design and to evaluate crew response in untested environments. It is to the benefit of any such model not only to collect relevant experimental data to correlate it against, but also to maintain an experimental standard or benchmark for future development in a readily and rapidly searchable and software accessible format. The Human thermal database project is intended to do just so; to collect relevant data from literature and experimentation and to store the data in a database structure for immediate and future use as a benchmark to judge human thermal models against, in identifying model strengths and weakness, to support model development and improve correlation, and to statistically quantify a model s predictive quality.

  6. Examining the Factors That Contribute to Successful Database Application Implementation Using the Technology Acceptance Model

    ERIC Educational Resources Information Center

    Nworji, Alexander O.

    2013-01-01

    Most organizations spend millions of dollars due to the impact of improperly implemented database application systems as evidenced by poor data quality problems. The purpose of this quantitative study was to use, and extend, the technology acceptance model (TAM) to assess the impact of information quality and technical quality factors on database…

  7. Spatial-temporal database model based on geodatabase

    NASA Astrophysics Data System (ADS)

    Zhu, Hongmei; Luo, Yu

    2009-10-01

    Entities in the real world have non-spatial attributes, as well as spatial and temporal features. A spatial-temporal data model aims at describing appropriately these intrinsic characteristics within the entities and model them on a conceptual level so that the model can present both static information and dynamic information that occurs over time. In this paper, we devise a novel spatial-temporal data model which is based on Geodatabase. The model employs object-oriented analysis method, combining object concept with event. The entity is defined as a feature class encapsulating attributes and operations. The operations detect change and store the changes automatically in a historic database in Geodatabase. Furthermore, the model takes advantage of the existing strengths of the relational database at the bottom level of Geodatabase, such as trigger and constraint, to monitor events on the attributes or locations and respond to the events correctly. A case of geographic database for Kunming municipal sewerage geographic information system is implemented by the model. The database reveals excellent performance on managing data and tracking the details of change. It provides a perfect data platform for querying, recurring history and predicting the trend of future. The instance demonstrates the spatial-temporal data model is efficient and practicable.

  8. 3MdB: the Mexican Million Models database

    NASA Astrophysics Data System (ADS)

    Morisset, C.; Delgado-Inglada, G.

    2014-10-01

    The 3MdB is an original effort to construct a large multipurpose database of photoionization models. This is a more modern version of a previous attempt based on Cloudy3D and IDL tools. It is accessed by MySQL requests. The models are obtained using the well known and widely used Cloudy photoionization code (Ferland et al, 2013). The database is aimed to host grids of models with different references to identify each project and to facilitate the extraction of the desired data. We present here a description of the way the database is managed and some of the projects that use 3MdB. Anybody can ask for a grid to be run and stored in 3MdB, to increase the visibility of the grid and the potential side applications of it.

  9. Materials Database Development for Ballistic Impact Modeling

    NASA Technical Reports Server (NTRS)

    Pereira, J. Michael

    2007-01-01

    A set of experimental data is being generated under the Fundamental Aeronautics Program Supersonics project to help create and validate accurate computational impact models of jet engine impact events. The data generated will include material property data generated at a range of different strain rates, from 1x10(exp -4)/sec to 5x10(exp 4)/sec, over a range of temperatures. In addition, carefully instrumented ballistic impact tests will be conducted on flat plates and curved structures to provide material and structural response information to help validate the computational models. The material property data and the ballistic impact data will be generated using materials from the same lot, as far as possible. It was found in preliminary testing that the surface finish of test specimens has an effect on measured high strain rate tension response of AL2024. Both the maximum stress and maximum elongation are greater on specimens with a smoother finish. This report gives an overview of the testing that is being conducted and presents results of preliminary testing of the surface finish study.

  10. Visual Analysis of Residuals from Data-Based Models in Complex Industrial Processes

    NASA Astrophysics Data System (ADS)

    Ordoñez, Daniel G.; Cuadrado, Abel A.; Díaz, Ignacio; García, Francisco J.; Díez, Alberto B.; Fuertes, Juan J.

    2012-10-01

    The use of data-based models for visualization purposes in an industrial background is discussed. Results using Self-Organizing Maps (SOM) show how through a good design of the model and a proper visualization of the residuals generated by the model itself, the behavior of essential parameters of the process can be easily tracked in a visual way. Real data from a cold rolling facility have been used to prove the advantages of these techniques.

  11. Database in low temperature plasma modeling

    NASA Astrophysics Data System (ADS)

    Sakai, Y.

    2002-05-01

    This article is composed of recommended sets of electron collision cross-sections and reaction cross-sections of excited species assessed by a swam method and of information on transport coefficients and reaction rates (cross-sections) of ions, which are needed in low temperature plasma modeling. These data have been piled up by the Investigation Committee on "Discharge Plasma Electron Collision Cross-sections", IEE Japan, and the author's laboratory. The gases taken for the assessment in this work are rare gases, Hg, N 2, O 2, CO 2, CF 4, CH 4, GeH 4, SiH 4, SF 6, C 2H 6, Si 2H 6, c-C 4F 8 and CCl 2F 2.

  12. Overarching framework for data-based modelling

    NASA Astrophysics Data System (ADS)

    Schelter, Björn; Mader, Malenka; Mader, Wolfgang; Sommerlade, Linda; Platt, Bettina; Lai, Ying-Cheng; Grebogi, Celso; Thiel, Marco

    2014-02-01

    One of the main modelling paradigms for complex physical systems are networks. When estimating the network structure from measured signals, typically several assumptions such as stationarity are made in the estimation process. Violating these assumptions renders standard analysis techniques fruitless. We here propose a framework to estimate the network structure from measurements of arbitrary non-linear, non-stationary, stochastic processes. To this end, we propose a rigorous mathematical theory that underlies this framework. Based on this theory, we present a highly efficient algorithm and the corresponding statistics that are immediately sensibly applicable to measured signals. We demonstrate its performance in a simulation study. In experiments of transitions between vigilance stages in rodents, we infer small network structures with complex, time-dependent interactions; this suggests biomarkers for such transitions, the key to understand and diagnose numerous diseases such as dementia. We argue that the suggested framework combines features that other approaches followed so far lack.

  13. BioProject and BioSample databases at NCBI: facilitating capture and organization of metadata

    PubMed Central

    Barrett, Tanya; Clark, Karen; Gevorgyan, Robert; Gorelenkov, Vyacheslav; Gribov, Eugene; Karsch-Mizrachi, Ilene; Kimelman, Michael; Pruitt, Kim D.; Resenchuk, Sergei; Tatusova, Tatiana; Yaschenko, Eugene; Ostell, James

    2012-01-01

    As the volume and complexity of data sets archived at NCBI grow rapidly, so does the need to gather and organize the associated metadata. Although metadata has been collected for some archival databases, previously, there was no centralized approach at NCBI for collecting this information and using it across databases. The BioProject database was recently established to facilitate organization and classification of project data submitted to NCBI, EBI and DDBJ databases. It captures descriptive information about research projects that result in high volume submissions to archival databases, ties together related data across multiple archives and serves as a central portal by which to inform users of data availability. Concomitantly, the BioSample database is being developed to capture descriptive information about the biological samples investigated in projects. BioProject and BioSample records link to corresponding data stored in archival repositories. Submissions are supported by a web-based Submission Portal that guides users through a series of forms for input of rich metadata describing their projects and samples. Together, these databases offer improved ways for users to query, locate, integrate and interpret the masses of data held in NCBI's archival repositories. The BioProject and BioSample databases are available at http://www.ncbi.nlm.nih.gov/bioproject and http://www.ncbi.nlm.nih.gov/biosample, respectively. PMID:22139929

  14. SPECTRAFACTORY.NET: A DATABASE OF MOLECULAR MODEL SPECTRA

    SciTech Connect

    Cami, J.; Van Malderen, R.; Markwick, A. J. E-mail: Andrew.Markwick@manchester.ac.uk

    2010-04-01

    We present a homogeneous database of synthetic molecular absorption and emission spectra from the optical to mm wavelengths for a large range of temperatures and column densities relevant for various astrophysical purposes, but in particular for the analysis, identification, and first-order analysis of molecular bands in spectroscopic observations. All spectra are calculated in the LTE limit from several molecular line lists, and are presented at various spectral resolving powers corresponding to several specific instrument simulations. The database is available online at http://www.spectrafactory.net, where users can freely browse, search, display, and download the spectra. We describe how additional model spectra can be requested for (automatic) calculation and inclusion. The database already contains over half a million model spectra for 39 molecules (96 different isotopologues) over the wavelength range 350 nm-3 mm ({approx}3-30000 cm{sup -1})

  15. An Approach to Query Cost Modelling in Numeric Databases.

    ERIC Educational Resources Information Center

    Jarvelin, Kalervo

    1989-01-01

    Examines factors that determine user charges based on query processing costs in numeric databases, and analyzes the problem of estimating such charges in advance. An approach to query cost estimation is presented which is based on the relational data model and the query optimization, cardinality estimation, and file design techniques developed in…

  16. Validating posttransplant hepatocellular carcinoma recurrence data in the United Network for Organ Sharing database.

    PubMed

    Samoylova, Mariya L; Dodge, Jennifer L; Vittinghoff, Eric; Yao, Francis Y; Roberts, John Paul

    2013-12-01

    The Organ Procurement and Transplantation Network (OPTN)/United Network for Organ Sharing (UNOS) database is the most comprehensive collection of liver transplantation data, but the quality of these data with respect to hepatocellular carcinoma (HCC) recurrence has not been well assessed. In this study, we compared observed HCC recurrence rates in the UNOS database to expected rates calculated with a hierarchical model for recurrence adjusted for recipient and tumor characteristics. We used the UNOS Standard Transplant Analysis and Research data set for adult transplant patients with an initial exception for an HCC diagnosis granted between January 1, 2006 and September 30, 2010 who underwent transplantation within the same time window. We developed a risk-adjusted Poisson model with patients as the unit of analysis, random effects for transplant centers, and years of follow-up as an offset to predict expected recurrences for each center. To further investigate the possibility of underreporting, we imputed expected recurrences for non-HCC deaths. In all, 5034 HCC liver transplant recipients were identified, and 6.8% experienced recurrence at a median of 1 year after transplantation. The covariate-adjusted shrinkage estimates of the observed/expected HCC recurrence ratios by transplant center ranged from 0.6 to 1.76 (median = 0.97). The 95% confidence intervals for the shrinkage ratios included unity for every center, and this indicated that none could be unambiguously identified as having lower or higher than expected HCC recurrence rates. Imputing outcomes for patients potentially experiencing unreported recurrence changed the center-specific shrinkage ratios to 0.72 to 1.39 (median = 0.98), with no centers having a shrinkage ratio significantly different from 1. The observed HCC recurrence rate was not significantly lower than the expected rate at any center, and this suggests that no systematic underreporting has occurred. This study validates the OPTN HCC

  17. Evaluating Service Organization Models

    PubMed Central

    TOUATI, NASSERA; PINEAULT, RAYNALD; CHAMPAGNE, FRANÇOIS; DENIS, JEAN-LOUIS; BROUSSELLE, ASTRID; CONTANDRIOPOULOS, ANDRÉ-PIERRE; GENEAU, ROBERT

    2016-01-01

    Based on the example of the evaluation of service organization models, this article shows how a configurational approach overcomes the limits of traditional methods which for the most part have studied the individual components of various models considered independently of one another. These traditional methods have led to results (observed effects) that are difficult to interpret. The configurational approach, in contrast, is based on the hypothesis that effects are associated with a set of internally coherent model features that form various configurations. These configurations, like their effects, are context-dependent. We explore the theoretical basis of the configuration approach in order to emphasize its relevance, and discuss the methodological challenges inherent in the application of this approach through an in-depth analysis of the scientific literature. We also propose methodological solutions to these challenges. We illustrate from an example how a configurational approach has been used to evaluate primary care models. Finally, we begin a discussion on the implications of this new evaluation approach for the scientific and decision-making communities. PMID:27274682

  18. Modeling past, current, and future time in medical databases.

    PubMed Central

    Kouramajian, V.; Fowler, J.

    1994-01-01

    Recent research has focused on increasing the power of medical information systems by incorporating time into the database system. A problem with much of this research is that it fails to differentiate between historical time and future time. The concept of bitemporal lifespan presented in this paper overcomes this deficiency. Bitemporal lifespan supports the concepts of valid time and transaction time and allows the integration of past, current, and future information in a unified model. The concept of bitemporal lifespan is presented within the framework of the Extended Entity-Relationship model. This model permits the characterization of temporal properties of entities, relationships, and attributes. Bitemporal constraints are defined that must hold between entities forming "isa" hierarchies and between entities and relationships. Finally, bitemporal extensions are presented for database query languages in order to provide natural high-level operators for bitemporal query expressions. PMID:7949941

  19. NGNP Risk Management Database: A Model for Managing Risk

    SciTech Connect

    John Collins

    2009-09-01

    To facilitate the implementation of the Risk Management Plan, the Next Generation Nuclear Plant (NGNP) Project has developed and employed an analytical software tool called the NGNP Risk Management System (RMS). A relational database developed in Microsoft® Access, the RMS provides conventional database utility including data maintenance, archiving, configuration control, and query ability. Additionally, the tool’s design provides a number of unique capabilities specifically designed to facilitate the development and execution of activities outlined in the Risk Management Plan. Specifically, the RMS provides the capability to establish the risk baseline, document and analyze the risk reduction plan, track the current risk reduction status, organize risks by reference configuration system, subsystem, and component (SSC) and Area, and increase the level of NGNP decision making.

  20. Artificial intelligence techniques for modeling database user behavior

    NASA Technical Reports Server (NTRS)

    Tanner, Steve; Graves, Sara J.

    1990-01-01

    The design and development of the adaptive modeling system is described. This system models how a user accesses a relational database management system in order to improve its performance by discovering use access patterns. In the current system, these patterns are used to improve the user interface and may be used to speed data retrieval, support query optimization and support a more flexible data representation. The system models both syntactic and semantic information about the user's access and employs both procedural and rule-based logic to manipulate the model.

  1. INTERCOMPARISON OF ALTERNATIVE VEGETATION DATABASES FOR REGIONAL AIR QUALITY MODELING

    EPA Science Inventory

    Vegetation cover data are used to characterize several regional air quality modeling processes, including the calculation of heat, moisture, and momentum fluxes with the Mesoscale Meteorological Model (MM5) and the estimate of biogenic volatile organic compound and nitric oxide...

  2. CyanOmics: an integrated database of omics for the model cyanobacterium Synechococcus sp. PCC 7002.

    PubMed

    Yang, Yaohua; Feng, Jie; Li, Tao; Ge, Feng; Zhao, Jindong

    2015-01-01

    Cyanobacteria are an important group of organisms that carry out oxygenic photosynthesis and play vital roles in both the carbon and nitrogen cycles of the Earth. The annotated genome of Synechococcus sp. PCC 7002, as an ideal model cyanobacterium, is available. A series of transcriptomic and proteomic studies of Synechococcus sp. PCC 7002 cells grown under different conditions have been reported. However, no database of such integrated omics studies has been constructed. Here we present CyanOmics, a database based on the results of Synechococcus sp. PCC 7002 omics studies. CyanOmics comprises one genomic dataset, 29 transcriptomic datasets and one proteomic dataset and should prove useful for systematic and comprehensive analysis of all those data. Powerful browsing and searching tools are integrated to help users directly access information of interest with enhanced visualization of the analytical results. Furthermore, Blast is included for sequence-based similarity searching and Cluster 3.0, as well as the R hclust function is provided for cluster analyses, to increase CyanOmics's usefulness. To the best of our knowledge, it is the first integrated omics analysis database for cyanobacteria. This database should further understanding of the transcriptional patterns, and proteomic profiling of Synechococcus sp. PCC 7002 and other cyanobacteria. Additionally, the entire database framework is applicable to any sequenced prokaryotic genome and could be applied to other integrated omics analysis projects. Database URL: http://lag.ihb.ac.cn/cyanomics.

  3. SAPling: a Scan-Add-Print barcoding database system to label and track asexual organisms

    PubMed Central

    Thomas, Michael A.; Schötz, Eva-Maria

    2011-01-01

    SUMMARY We have developed a ‘Scan-Add-Print’ database system, SAPling, to track and monitor asexually reproducing organisms. Using barcodes to uniquely identify each animal, we can record information on the life of the individual in a computerized database containing its entire family tree. SAPling has enabled us to carry out large-scale population dynamics experiments with thousands of planarians and keep track of each individual. The database stores information such as family connections, birth date, division date and generation. We show that SAPling can be easily adapted to other asexually reproducing organisms and has a strong potential for use in large-scale and/or long-term population and senescence studies as well as studies of clonal diversity. The software is platform-independent, designed for reliability and ease of use, and provided open source from our webpage to allow project-specific customization. PMID:21993779

  4. Applications of the Cambridge Structural Database in organic chemistry and crystal chemistry.

    PubMed

    Allen, Frank H; Motherwell, W D Samuel

    2002-06-01

    The Cambridge Structural Database (CSD) and its associated software systems have formed the basis for more than 800 research applications in structural chemistry, crystallography and the life sciences. Relevant references, dating from the mid-1970s, and brief synopses of these papers are collected in a database, DBUse, which is freely available via the CCDC website. This database has been used to review research applications of the CSD in organic chemistry, including supramolecular applications, and in organic crystal chemistry. The review concentrates on applications that have been published since 1990 and covers a wide range of topics, including structure correlation, conformational analysis, hydrogen bonding and other intermolecular interactions, studies of crystal packing, extended structural motifs, crystal engineering and polymorphism, and crystal structure prediction. Applications of CSD information in studies of crystal structure precision, the determination of crystal structures from powder diffraction data, together with applications in chemical informatics, are also discussed.

  5. Standards, databases, and modeling tools in systems biology.

    PubMed

    Kohl, Michael

    2011-01-01

    Modeling is a means for integrating the results from Genomics, Transcriptomics, Proteomics, and Metabolomics experiments and for gaining insights into the interaction of the constituents of biological systems. However, sharing such large amounts of frequently heterogeneous and distributed experimental data needs both standard data formats and public repositories. Standardization and a public storage system are also important for modeling due to the possibility of sharing models irrespective of the used software tools. Furthermore, rapid model development strongly benefits from available software packages that relieve the modeler of recurring tasks like numerical integration of rate equations or parameter estimation. In this chapter, the most common standard formats used for model encoding and some of the major public databases in this scientific field are presented. The main features of currently available modeling software are discussed and proposals for the application of such tools are given.

  6. ASGARD: an open-access database of annotated transcriptomes for emerging model arthropod species.

    PubMed

    Zeng, Victor; Extavour, Cassandra G

    2012-01-01

    The increased throughput and decreased cost of next-generation sequencing (NGS) have shifted the bottleneck genomic research from sequencing to annotation, analysis and accessibility. This is particularly challenging for research communities working on organisms that lack the basic infrastructure of a sequenced genome, or an efficient way to utilize whatever sequence data may be available. Here we present a new database, the Assembled Searchable Giant Arthropod Read Database (ASGARD). This database is a repository and search engine for transcriptomic data from arthropods that are of high interest to multiple research communities but currently lack sequenced genomes. We demonstrate the functionality and utility of ASGARD using de novo assembled transcriptomes from the milkweed bug Oncopeltus fasciatus, the cricket Gryllus bimaculatus and the amphipod crustacean Parhyale hawaiensis. We have annotated these transcriptomes to assign putative orthology, coding region determination, protein domain identification and Gene Ontology (GO) term annotation to all possible assembly products. ASGARD allows users to search all assemblies by orthology annotation, GO term annotation or Basic Local Alignment Search Tool. User-friendly features of ASGARD include search term auto-completion suggestions based on database content, the ability to download assembly product sequences in FASTA format, direct links to NCBI data for predicted orthologs and graphical representation of the location of protein domains and matches to similar sequences from the NCBI non-redundant database. ASGARD will be a useful repository for transcriptome data from future NGS studies on these and other emerging model arthropods, regardless of sequencing platform, assembly or annotation status. This database thus provides easy, one-stop access to multi-species annotated transcriptome information. We anticipate that this database will be useful for members of multiple research communities, including developmental

  7. ASGARD: an open-access database of annotated transcriptomes for emerging model arthropod species.

    PubMed

    Zeng, Victor; Extavour, Cassandra G

    2012-01-01

    The increased throughput and decreased cost of next-generation sequencing (NGS) have shifted the bottleneck genomic research from sequencing to annotation, analysis and accessibility. This is particularly challenging for research communities working on organisms that lack the basic infrastructure of a sequenced genome, or an efficient way to utilize whatever sequence data may be available. Here we present a new database, the Assembled Searchable Giant Arthropod Read Database (ASGARD). This database is a repository and search engine for transcriptomic data from arthropods that are of high interest to multiple research communities but currently lack sequenced genomes. We demonstrate the functionality and utility of ASGARD using de novo assembled transcriptomes from the milkweed bug Oncopeltus fasciatus, the cricket Gryllus bimaculatus and the amphipod crustacean Parhyale hawaiensis. We have annotated these transcriptomes to assign putative orthology, coding region determination, protein domain identification and Gene Ontology (GO) term annotation to all possible assembly products. ASGARD allows users to search all assemblies by orthology annotation, GO term annotation or Basic Local Alignment Search Tool. User-friendly features of ASGARD include search term auto-completion suggestions based on database content, the ability to download assembly product sequences in FASTA format, direct links to NCBI data for predicted orthologs and graphical representation of the location of protein domains and matches to similar sequences from the NCBI non-redundant database. ASGARD will be a useful repository for transcriptome data from future NGS studies on these and other emerging model arthropods, regardless of sequencing platform, assembly or annotation status. This database thus provides easy, one-stop access to multi-species annotated transcriptome information. We anticipate that this database will be useful for members of multiple research communities, including developmental

  8. The Medicago Genome Initiative: a model legume database

    PubMed Central

    Bell, Callum J.; Dixon, Richard A.; Farmer, Andrew D.; Flores, Raul; Inman, Jeff; Gonzales, Robert A.; Harrison, Maria J.; Paiva, Nancy L.; Scott, Angela D.; Weller, Jennifer W.; May, Gregory D.

    2001-01-01

    The Medicago Genome Initiative (MGI) is a database of EST sequences of the model legume Medicago truncatula. The database is available to the public and has resulted from a collaborative research effort between the Samuel Roberts Noble Foundation and the National Center for Genome Resources to investigate the genome of M.truncatula. MGI is part of the greater integrated Medicago functional genomics program at the Noble Foundation (http://www.noble .org), which is taking a global approach in studying the genetic and biochemical events associated with the growth, development and environmental interactions of this model legume. Our approach will include: large-scale EST sequencing, gene expression profiling, the generation of M.truncatula activation-tagged and promoter trap insertion mutants, high-throughput metabolic profiling, and proteome studies. These multidisciplinary information pools will be interfaced with one another to provide scientists with an integrated, holistic set of tools to address fundamental questions pertaining to legume biology. The public interface to the MGI database can be accessed at http://www.ncgr.org/research/mgi. PMID:11125064

  9. Databases, models, and algorithms for functional genomics: a bioinformatics perspective.

    PubMed

    Singh, Gautam B; Singh, Harkirat

    2005-02-01

    A variety of patterns have been observed on the DNA and protein sequences that serve as control points for gene expression and cellular functions. Owing to the vital role of such patterns discovered on biological sequences, they are generally cataloged and maintained within internationally shared databases. Furthermore,the variability in a family of observed patterns is often represented using computational models in order to facilitate their search within an uncharacterized biological sequence. As the biological data is comprised of a mosaic of sequence-levels motifs, it is significant to unravel the synergies of macromolecular coordination utilized in cell-specific differential synthesis of proteins. This article provides an overview of the various pattern representation methodologies and the surveys the pattern databases available for use to the molecular biologists. Our aim is to describe the principles behind the computational modeling and analysis techniques utilized in bioinformatics research, with the objective of providing insight necessary to better understand and effectively utilize the available databases and analysis tools. We also provide a detailed review of DNA sequence level patterns responsible for structural conformations within the Scaffold or Matrix Attachment Regions (S/MARs).

  10. NGNP Risk Management Database: A Model for Managing Risk

    SciTech Connect

    John Collins; John M. Beck

    2011-11-01

    The Next Generation Nuclear Plant (NGNP) Risk Management System (RMS) is a database used to maintain the project risk register. The RMS also maps risk reduction activities to specific identified risks. Further functionality of the RMS includes mapping reactor suppliers Design Data Needs (DDNs) to risk reduction tasks and mapping Phenomena Identification Ranking Table (PIRTs) to associated risks. This document outlines the basic instructions on how to use the RMS. This document constitutes Revision 1 of the NGNP Risk Management Database: A Model for Managing Risk. It incorporates the latest enhancements to the RMS. The enhancements include six new custom views of risk data - Impact/Consequence, Tasks by Project Phase, Tasks by Status, Tasks by Project Phase/Status, Tasks by Impact/WBS, and Tasks by Phase/Impact/WBS.

  11. Mouse Genome Database: From sequence to phenotypes and disease models.

    PubMed

    Eppig, Janan T; Richardson, Joel E; Kadin, James A; Smith, Cynthia L; Blake, Judith A; Bult, Carol J

    2015-08-01

    The Mouse Genome Database (MGD, www.informatics.jax.org) is the international scientific database for genetic, genomic, and biological data on the laboratory mouse to support the research requirements of the biomedical community. To accomplish this goal, MGD provides broad data coverage, serves as the authoritative standard for mouse nomenclature for genes, mutants, and strains, and curates and integrates many types of data from literature and electronic sources. Among the key data sets MGD supports are: the complete catalog of mouse genes and genome features, comparative homology data for mouse and vertebrate genes, the authoritative set of Gene Ontology (GO) annotations for mouse gene functions, a comprehensive catalog of mouse mutations and their phenotypes, and a curated compendium of mouse models of human diseases. Here, we describe the data acquisition process, specifics about MGD's key data areas, methods to access and query MGD data, and outreach and user help facilities. PMID:26150326

  12. PANTHER: a browsable database of gene products organized by biological function, using curated protein family and subfamily classification.

    PubMed

    Thomas, Paul D; Kejariwal, Anish; Campbell, Michael J; Mi, Huaiyu; Diemer, Karen; Guo, Nan; Ladunga, Istvan; Ulitsky-Lazareva, Betty; Muruganujan, Anushya; Rabkin, Steven; Vandergriff, Jody A; Doremieux, Olivier

    2003-01-01

    The PANTHER database was designed for high-throughput analysis of protein sequences. One of the key features is a simplified ontology of protein function, which allows browsing of the database by biological functions. Biologist curators have associated the ontology terms with groups of protein sequences rather than individual sequences. Statistical models (Hidden Markov Models, or HMMs) are built from each of these groups. The advantage of this approach is that new sequences can be automatically classified as they become available. To ensure accurate functional classification, HMMs are constructed not only for families, but also for functionally distinct subfamilies. Multiple sequence alignments and phylogenetic trees, including curator-assigned information, are available for each family. The current version of the PANTHER database includes training sequences from all organisms in the GenBank non-redundant protein database, and the HMMs have been used to classify gene products across the entire genomes of human, and Drosophila melanogaster. The ontology terms and protein families and subfamilies, as well as Drosophila gene c;assifications, can be browsed and searched for free. Due to outstanding contractual obligations, access to human gene classifications and to protein family trees and multiple sequence alignments will temporarily require a nominal registration fee. PANTHER is publicly available on the web at http://panther.celera.com.

  13. Organizing the Extremely Large LSST Database forReal-Time Astronomical Processing

    SciTech Connect

    Becla, Jacek; Lim, Kian-Tat; Monkewitz, Serge; Nieto-Santisteban, Maria; Thakar, Ani; /Johns Hopkins U.

    2007-11-07

    The Large Synoptic Survey Telescope (LSST) will catalog billions of astronomical objects and trillions of sources, all of which will be stored and managed by a database management system. One of the main challenges is real-time alert generation. To generate alerts, up to 100K new difference detections have to be cross-correlated with the huge historical catalogs, and then further processed to prune false alerts. This paper explains the challenges, the implementation of the LSST Association Pipeline and the database organization strategies we are planning to use to meet the real-time requirements, including data partitioning, parallelization, and pre-loading.

  14. Organization's Orderly Interest Exploration: Inception, Development and Insights of AIAA's Topics Database

    NASA Technical Reports Server (NTRS)

    Marshall, Jospeh R.; Morris, Allan T.

    2007-01-01

    Since 2003, AIAA's Computer Systems and Software Systems Technical Committees (TCs) have developed a database that aids technical committee management to map technical topics to their members. This Topics/Interest (T/I) database grew out of a collection of charts and spreadsheets maintained by the TCs. Since its inception, the tool has evolved into a multi-dimensional database whose dimensions include the importance, interest and expertise of TC members and whether or not a member and/or a TC is actively involved with the topic. In 2005, the database was expanded to include the TCs in AIAA s Information Systems Group and then expanded further to include all AIAA TCs. It was field tested at an AIAA Technical Activities Committee (TAC) Workshop in early 2006 through live access by over 80 users. Through the use of the topics database, TC and program committee (PC) members can accomplish relevant tasks such as: to identify topic experts (for Aerospace America articles or external contacts), to determine the interest of its members, to identify overlapping topics between diverse TCs and PCs, to guide new member drives and to reveal emerging topics. This paper will describe the origins, inception, initial development, field test and current version of the tool as well as elucidate the benefits and insights gained by using the database to aid the management of various TC functions. Suggestions will be provided to guide future development of the database for the purpose of providing dynamics and system level benefits to AIAA that currently do not exist in any technical organization.

  15. Teaching biology with model organisms

    NASA Astrophysics Data System (ADS)

    Keeley, Dolores A.

    The purpose of this study is to identify and use model organisms that represent each of the kingdoms biologists use to classify organisms, while experiencing the process of science through guided inquiry. The model organisms will be the basis for studying the four high school life science core ideas as identified by the Next Generation Science Standards (NGSS): LS1-From molecules to organisms, LS2-Ecosystems, LS3- Heredity, and LS4- Biological Evolution. NGSS also have identified four categories of science and engineering practices which include developing and using models and planning and carrying out investigations. The living organisms will be utilized to increase student interest and knowledge within the discipline of Biology. Pre-test and posttest analysis utilizing student t-test analysis supported the hypothesis. This study shows increased student learning as a result of using living organisms as models for classification and working in an inquiry-based learning environment.

  16. Feasibility and utility of applications of the common data model to multiple, disparate observational health databases

    PubMed Central

    Makadia, Rupa; Matcho, Amy; Ma, Qianli; Knoll, Chris; Schuemie, Martijn; DeFalco, Frank J; Londhe, Ajit; Zhu, Vivienne; Ryan, Patrick B

    2015-01-01

    Objectives To evaluate the utility of applying the Observational Medical Outcomes Partnership (OMOP) Common Data Model (CDM) across multiple observational databases within an organization and to apply standardized analytics tools for conducting observational research. Materials and methods Six deidentified patient-level datasets were transformed to the OMOP CDM. We evaluated the extent of information loss that occurred through the standardization process. We developed a standardized analytic tool to replicate the cohort construction process from a published epidemiology protocol and applied the analysis to all 6 databases to assess time-to-execution and comparability of results. Results Transformation to the CDM resulted in minimal information loss across all 6 databases. Patients and observations excluded were due to identified data quality issues in the source system, 96% to 99% of condition records and 90% to 99% of drug records were successfully mapped into the CDM using the standard vocabulary. The full cohort replication and descriptive baseline summary was executed for 2 cohorts in 6 databases in less than 1 hour. Discussion The standardization process improved data quality, increased efficiency, and facilitated cross-database comparisons to support a more systematic approach to observational research. Comparisons across data sources showed consistency in the impact of inclusion criteria, using the protocol and identified differences in patient characteristics and coding practices across databases. Conclusion Standardizing data structure (through a CDM), content (through a standard vocabulary with source code mappings), and analytics can enable an institution to apply a network-based approach to observational research across multiple, disparate observational health databases. PMID:25670757

  17. An atmospheric tritium release database for model comparisons

    SciTech Connect

    Murphy, C.E. Jr.; Wortham, G.R.

    1991-12-19

    A database of vegetation, soil, and air tritium concentrations at gridded coordinate locations following nine accidental atmospheric releases is described. While none of the releases caused a significant dose to the public, the data collected is valuable for comparison with the results of tritium transport models used for risk assessment. The largest, potential, individual off-site dose from any of the releases was calculated to be 1.6 mrem. The population dose from this same release was 46 person-rem which represents 0.04% of the natural background radiation dose to the population in the path of the release.

  18. Database and Interim Glass Property Models for Hanford HLW Glasses

    SciTech Connect

    Hrma, Pavel R.; Piepel, Gregory F.; Vienna, John D.; Cooley, Scott K.; Kim, Dong-Sang; Russell, Renee L.

    2001-07-24

    The purpose of this report is to provide a methodology for an increase in the efficiency and a decrease in the cost of vitrifying high-level waste (HLW) by optimizing HLW glass formulation. This methodology consists in collecting and generating a database of glass properties that determine HLW glass processability and acceptability and relating these properties to glass composition. The report explains how the property-composition models are developed, fitted to data, used for glass formulation optimization, and continuously updated in response to changes in HLW composition estimates and changes in glass processing technology. Further, the report reviews the glass property-composition literature data and presents their preliminary critical evaluation and screening. Finally the report provides interim property-composition models for melt viscosity, for liquidus temperature (with spinel and zircon primary crystalline phases), and for the product consistency test normalized releases of B, Na, and Li. Models were fitted to a subset of the screened database deemed most relevant for the current HLW composition region.

  19. Modeling, Measurements, and Fundamental Database Development for Nonequilibrium Hypersonic Aerothermodynamics

    NASA Technical Reports Server (NTRS)

    Bose, Deepak

    2012-01-01

    The design of entry vehicles requires predictions of aerothermal environment during the hypersonic phase of their flight trajectories. These predictions are made using computational fluid dynamics (CFD) codes that often rely on physics and chemistry models of nonequilibrium processes. The primary processes of interest are gas phase chemistry, internal energy relaxation, electronic excitation, nonequilibrium emission and absorption of radiation, and gas-surface interaction leading to surface recession and catalytic recombination. NASAs Hypersonics Project is advancing the state-of-the-art in modeling of nonequilibrium phenomena by making detailed spectroscopic measurements in shock tube and arcjets, using ab-initio quantum mechanical techniques develop fundamental chemistry and spectroscopic databases, making fundamental measurements of finite-rate gas surface interactions, implementing of detailed mechanisms in the state-of-the-art CFD codes, The development of new models is based on validation with relevant experiments. We will present the latest developments and a roadmap for the technical areas mentioned above

  20. Developing High-resolution Soil Database for Regional Crop Modeling in East Africa

    NASA Astrophysics Data System (ADS)

    Han, E.; Ines, A. V. M.

    2014-12-01

    The most readily available soil data for regional crop modeling in Africa is the World Inventory of Soil Emission potentials (WISE) dataset, which has 1125 soil profiles for the world, but does not extensively cover countries Ethiopia, Kenya, Uganda and Tanzania in East Africa. Another dataset available is the HC27 (Harvest Choice by IFPRI) in a gridded format (10km) but composed of generic soil profiles based on only three criteria (texture, rooting depth, and organic carbon content). In this paper, we present a development and application of a high-resolution (1km), gridded soil database for regional crop modeling in East Africa. Basic soil information is extracted from Africa Soil Information Service (AfSIS), which provides essential soil properties (bulk density, soil organic carbon, soil PH and percentages of sand, silt and clay) for 6 different standardized soil layers (5, 15, 30, 60, 100 and 200 cm) in 1km resolution. Soil hydraulic properties (e.g., field capacity and wilting point) are derived from the AfSIS soil dataset using well-proven pedo-transfer functions and are customized for DSSAT-CSM soil data requirements. The crop model is used to evaluate crop yield forecasts using the new high resolution soil database and compared with WISE and HC27. In this paper we will present also the results of DSSAT loosely coupled with a hydrologic model (VIC) to assimilate root-zone soil moisture. Creating a grid-based soil database, which provides a consistent soil input for two different models (DSSAT and VIC) is a critical part of this work. The created soil database is expected to contribute to future applications of DSSAT crop simulation in East Africa where food security is highly vulnerable.

  1. Models and Results Database (MAR-D), Version 4. 0

    SciTech Connect

    Branham-Haar, K.A.; Dinneen, R.A.; Russell, K.D.; Skinner, N.L. )

    1992-05-01

    The Nuclear Regulatory Commission's Office of Nuclear Regulatory Research (NRC-RES) is presently funding the development of the Models and Results Database (MAR-D) at the Idaho National Engineering Laboratory. MAR-D's primary function is to create a data repository for NUREG-1150 and other permanent data by providing input, conversion, and output capabilities for data used by IRRAS, SARA, SETS, and FRANTIC personal computer (PC) codes. As probabilistic risk assessments and individual plant examinations are submitted to the NRC for review, MAR-D can be used to convert the models and results from the study for use with IRRAS and SARA. Then, these data can be easily accessed by future studies and will be in a form that will enhance the analysis process. This reference manual provides an overview of the function available within MAR-D and step-by-step operating instructions.

  2. New Asteroid Shape Models Derived from the Lowell Photometric Database

    NASA Astrophysics Data System (ADS)

    Durech, Josef; Hanus, J.; Vanco, R.; Oszkiewicz, D.; Bowell, E.

    2013-10-01

    Asteroid shapes and spin states can be derived from their disk-integrated sparse-in-time photometry by the lightcurve inversion method. A huge amount of sparse photometry is provided by current all-sky surveys. However, the data from surveys suffer from large random and systematic errors. Oszkiewicz et al. (2011, JQSRT 112, 1919) partly removed the systematic trends in the photometry reported to the MPC and created the so-called 'Lowell photometric database'. The database consists of re-calibrated photometry for about 500,000 asteroids, with typically hundreds of brightness measurements per object. Bowell et al. (M&PS, submitted) used this database to analyze brightness variations with ecliptic longitude and estimated spin-axis longitudes for about 350,000 asteroids. In our work, we processed data for the first 10,000 numbered asteroids with the lightcurve inversion method (Kaasalainen et al., 2001, Icarus 153,37) using an enormous computational power of Asteroids@home (http://asteroidsathome.net) - a distributed computing project built on the BOINC platform. More than 10,000 users have joined the project and their computers were used for the time-consuming search for the sidereal rotation period in the sparse data. Although the photometric accuracy of the Lowell data is low 0.2 mag), we were able to find unique models for several hundred asteroids. We will present the first results based on the statistical analysis of the sample (distribution of spin vectors, for example) and we will also discuss the relevance of our approach to Gaia, LSST, ATLAS, and other future sources of asteroid photometry with sparse sampling.

  3. MetRxn: a knowledgebase of metabolites and reactions spanning metabolic models and databases

    PubMed Central

    2012-01-01

    Background Increasingly, metabolite and reaction information is organized in the form of genome-scale metabolic reconstructions that describe the reaction stoichiometry, directionality, and gene to protein to reaction associations. A key bottleneck in the pace of reconstruction of new, high-quality metabolic models is the inability to directly make use of metabolite/reaction information from biological databases or other models due to incompatibilities in content representation (i.e., metabolites with multiple names across databases and models), stoichiometric errors such as elemental or charge imbalances, and incomplete atomistic detail (e.g., use of generic R-group or non-explicit specification of stereo-specificity). Description MetRxn is a knowledgebase that includes standardized metabolite and reaction descriptions by integrating information from BRENDA, KEGG, MetaCyc, Reactome.org and 44 metabolic models into a single unified data set. All metabolite entries have matched synonyms, resolved protonation states, and are linked to unique structures. All reaction entries are elementally and charge balanced. This is accomplished through the use of a workflow of lexicographic, phonetic, and structural comparison algorithms. MetRxn allows for the download of standardized versions of existing genome-scale metabolic models and the use of metabolic information for the rapid reconstruction of new ones. Conclusions The standardization in description allows for the direct comparison of the metabolite and reaction content between metabolic models and databases and the exhaustive prospecting of pathways for biotechnological production. This ever-growing dataset currently consists of over 76,000 metabolites participating in more than 72,000 reactions (including unresolved entries). MetRxn is hosted on a web-based platform that uses relational database models (MySQL). PMID:22233419

  4. Using chemical organization theory for model checking

    PubMed Central

    Kaleta, Christoph; Richter, Stephan; Dittrich, Peter

    2009-01-01

    Motivation: The increasing number and complexity of biomodels makes automatic procedures for checking the models' properties and quality necessary. Approaches like elementary mode analysis, flux balance analysis, deficiency analysis and chemical organization theory (OT) require only the stoichiometric structure of the reaction network for derivation of valuable information. In formalisms like Systems Biology Markup Language (SBML), however, information about the stoichiometric coefficients required for an analysis of chemical organizations can be hidden in kinetic laws. Results: First, we introduce an algorithm that uncovers stoichiometric information that might be hidden in the kinetic laws of a reaction network. This allows us to apply OT to SBML models using modifiers. Second, using the new algorithm, we performed a large-scale analysis of the 185 models contained in the manually curated BioModels Database. We found that for 41 models (22%) the set of organizations changes when modifiers are considered correctly. We discuss one of these models in detail (BIOMD149, a combined model of the ERK- and Wnt-signaling pathways), whose set of organizations drastically changes when modifiers are considered. Third, we found inconsistencies in 5 models (3%) and identified their characteristics. Compared with flux-based methods, OT is able to identify those species and reactions more accurately [in 26 cases (14%)] that can be present in a long-term simulation of the model. We conclude that our approach is a valuable tool that helps to improve the consistency of biomodels and their repositories. Availability: All data and a JAVA applet to check SBML-models is available from http://www.minet.uni-jena.de/csb/prj/ot/tools Contact: dittrich@minet.uni-jena.de Supplementary information: Supplementary data are available at Bioinformatics online. PMID:19468053

  5. Data-based Non-Markovian Model Inference

    NASA Astrophysics Data System (ADS)

    Ghil, Michael

    2015-04-01

    This talk concentrates on obtaining stable and efficient data-based models for simulation and prediction in the geosciences and life sciences. The proposed model derivation relies on using a multivariate time series of partial observations from a large-dimensional system, and the resulting low-order models are compared with the optimal closures predicted by the non-Markovian Mori-Zwanzig formalism of statistical physics. Multilayer stochastic models (MSMs) are introduced as both a very broad generalization and a time-continuous limit of existing multilevel, regression-based approaches to data-based closure, in particular of empirical model reduction (EMR). We show that the multilayer structure of MSMs can provide a natural Markov approximation to the generalized Langevin equation (GLE) of the Mori-Zwanzig formalism. A simple correlation-based stopping criterion for an EMR-MSM model is derived to assess how well it approximates the GLE solution. Sufficient conditions are given for the nonlinear cross-interactions between the constitutive layers of a given MSM to guarantee the existence of a global random attractor. This existence ensures that no blow-up can occur for a very broad class of MSM applications. The EMR-MSM methodology is first applied to a conceptual, nonlinear, stochastic climate model of coupled slow and fast variables, in which only slow variables are observed. The resulting reduced model with energy-conserving nonlinearities captures the main statistical features of the slow variables, even when there is no formal scale separation and the fast variables are quite energetic. Second, an MSM is shown to successfully reproduce the statistics of a partially observed, generalized Lokta-Volterra model of population dynamics in its chaotic regime. The positivity constraint on the solutions' components replaces here the quadratic-energy-preserving constraint of fluid-flow problems and it successfully prevents blow-up. This work is based on a close

  6. MOSAIC: An organic geochemical and sedimentological database for marine surface sediments

    NASA Astrophysics Data System (ADS)

    Tavagna, Maria Luisa; Usman, Muhammed; De Avelar, Silvania; Eglinton, Timothy

    2015-04-01

    Modern ocean sediments serve as the interface between the biosphere and the geosphere, play a key role in biogeochemical cycles and provide a window on how contemporary processes are written into the sedimentary record. Research over past decades has resulted in a wealth of information on the content and composition of organic matter in marine sediments, with ever-more sophisticated techniques continuing to yield information of greater detail and as an accelerating pace. However, there has been no attempt to synthesize this wealth of information. We are establishing a new database that incorporates information relevant to local, regional and global-scale assessment of the content, source and fate of organic materials accumulating in contemporary marine sediments. In the MOSAIC (Modern Ocean Sediment Archive and Inventory of Carbon) database, particular emphasis is placed on molecular and isotopic information, coupled with relevant contextual information (e.g., sedimentological properties) relevant to elucidating factors that influence the efficiency and nature of organic matter burial. The main features of MOSAIC include: (i) Emphasis on continental margin sediments as major loci of carbon burial, and as the interface between terrestrial and oceanic realms; (ii) Bulk to molecular-level organic geochemical properties and parameters, including concentration and isotopic compositions; (iii) Inclusion of extensive contextual data regarding the depositional setting, in particular with respect to sedimentological and redox characteristics. The ultimate goal is to create an open-access instrument, available on the web, to be utilized for research and education by the international community who can both contribute to, and interrogate the database. The submission will be accomplished by means of a pre-configured table available on the MOSAIC webpage. The information on the filled tables will be checked and eventually imported, via the Structural Query Language (SQL), into

  7. LANL High-Level Model (HLM) database development letter report

    SciTech Connect

    1995-10-01

    Traditional methods of evaluating munitions have been able to successfully compare like munition`s capabilities. On the modern battlefield, however, many different types of munitions compete for the same set of targets. Assessing the overall stockpile capability and proper mix of these weapons is not a simple task, as their use depends upon the specific geographic region of the world, the threat capabilities, the tactics and operational strategy used by both the US and Threat commanders, and of course the type and quantity of munitions available to the CINC. To sort out these types of issues, a hierarchical set of dynamic, two-sided combat simulations are generally used. The DoD has numerous suitable models for this purpose, but rarely are the models focused on munitions expenditures. Rather, they are designed to perform overall platform assessments and force mix evaluations. However, in some cases, the models could be easily adapted to provide this information, since it is resident in the model`s database. Unfortunately, these simulations` complexity (their greatest strength) precludes quick turnaround assessments of the type and scope required by senior decision-makers.

  8. Global search tool for the Advanced Photon Source Integrated Relational Model of Installed Systems (IRMIS) database.

    SciTech Connect

    Quock, D. E. R.; Cianciarulo, M. B.; APS Engineering Support Division; Purdue Univ.

    2007-01-01

    The Integrated Relational Model of Installed Systems (IRMIS) is a relational database tool that has been implemented at the Advanced Photon Source to maintain an updated account of approximately 600 control system software applications, 400,000 process variables, and 30,000 control system hardware components. To effectively display this large amount of control system information to operators and engineers, IRMIS was initially built with nine Web-based viewers: Applications Organizing Index, IOC, PLC, Component Type, Installed Components, Network, Controls Spares, Process Variables, and Cables. However, since each viewer is designed to provide details from only one major category of the control system, the necessity for a one-stop global search tool for the entire database became apparent. The user requirements for extremely fast database search time and ease of navigation through search results led to the choice of Asynchronous JavaScript and XML (AJAX) technology in the implementation of the IRMIS global search tool. Unique features of the global search tool include a two-tier level of displayed search results, and a database data integrity validation and reporting mechanism.

  9. A future of the model organism model

    PubMed Central

    Rine, Jasper

    2014-01-01

    Changes in technology are fundamentally reframing our concept of what constitutes a model organism. Nevertheless, research advances in the more traditional model organisms have enabled fresh and exciting opportunities for young scientists to establish new careers and offer the hope of comprehensive understanding of fundamental processes in life. New advances in translational research can be expected to heighten the importance of basic research in model organisms and expand opportunities. However, researchers must take special care and implement new resources to enable the newest members of the community to engage fully with the remarkable legacy of information in these fields. PMID:24577733

  10. An atmospheric tritium release database for model comparisons

    SciTech Connect

    Murphy, C.E. Jr.; Wortham, G.R.

    1997-10-13

    A database of vegetation, soil, and air tritium concentrations at gridded coordinate locations following nine accidental atmospheric releases is described. The concentration data is supported by climatological data taken during and immediately after the releases. In six cases, the release data is supplemented with meteorological data taken at seven towers scattered throughout the immediate area of the releases and data from a single television tower instrumented at eight heights. While none of the releases caused a significant dose to the public, the data collected is valuable for comparison with the results of tritium transport models used for risk assessment. The largest, potential off-site dose from any of the releases was calculated to be 1.6 mrem. The population dose from this same release was 46 person-rem which represents 0.04 percent of the natural background dose to the population in the path of the release.

  11. BioModels Database: a repository of mathematical models of biological processes.

    PubMed

    Chelliah, Vijayalakshmi; Laibe, Camille; Le Novère, Nicolas

    2013-01-01

    BioModels Database is a public online resource that allows storing and sharing of published, peer-reviewed quantitative, dynamic models of biological processes. The model components and behaviour are thoroughly checked to correspond the original publication and manually curated to ensure reliability. Furthermore, the model elements are annotated with terms from controlled vocabularies as well as linked to relevant external data resources. This greatly helps in model interpretation and reuse. Models are stored in SBML format, accepted in SBML and CellML formats, and are available for download in various other common formats such as BioPAX, Octave, SciLab, VCML, XPP and PDF, in addition to SBML. The reaction network diagram of the models is also available in several formats. BioModels Database features a search engine, which provides simple and more advanced searches. Features such as online simulation and creation of smaller models (submodels) from the selected model elements of a larger one are provided. BioModels Database can be accessed both via a web interface and programmatically via web services. New models are available in BioModels Database at regular releases, about every 4 months. PMID:23715986

  12. BioModels Database: An enhanced, curated and annotated resource for published quantitative kinetic models

    PubMed Central

    2010-01-01

    Background Quantitative models of biochemical and cellular systems are used to answer a variety of questions in the biological sciences. The number of published quantitative models is growing steadily thanks to increasing interest in the use of models as well as the development of improved software systems and the availability of better, cheaper computer hardware. To maximise the benefits of this growing body of models, the field needs centralised model repositories that will encourage, facilitate and promote model dissemination and reuse. Ideally, the models stored in these repositories should be extensively tested and encoded in community-supported and standardised formats. In addition, the models and their components should be cross-referenced with other resources in order to allow their unambiguous identification. Description BioModels Database http://www.ebi.ac.uk/biomodels/ is aimed at addressing exactly these needs. It is a freely-accessible online resource for storing, viewing, retrieving, and analysing published, peer-reviewed quantitative models of biochemical and cellular systems. The structure and behaviour of each simulation model distributed by BioModels Database are thoroughly checked; in addition, model elements are annotated with terms from controlled vocabularies as well as linked to relevant data resources. Models can be examined online or downloaded in various formats. Reaction network diagrams generated from the models are also available in several formats. BioModels Database also provides features such as online simulation and the extraction of components from large scale models into smaller submodels. Finally, the system provides a range of web services that external software systems can use to access up-to-date data from the database. Conclusions BioModels Database has become a recognised reference resource for systems biology. It is being used by the community in a variety of ways; for example, it is used to benchmark different simulation

  13. Podiform chromite deposits--database and grade and tonnage models

    USGS Publications Warehouse

    Mosier, Dan L.; Singer, Donald A.; Moring, Barry C.; Galloway, John P.

    2012-01-01

    Chromite ((Mg, Fe++)(Cr, Al, Fe+++)2O4) is the only source for the metallic element chromium, which is used in the metallurgical, chemical, and refractory industries. Podiform chromite deposits are small magmatic chromite bodies formed in the ultramafic section of an ophiolite complex in the oceanic crust. These deposits have been found in midoceanic ridge, off-ridge, and suprasubduction tectonic settings. Most podiform chromite deposits are found in dunite or peridotite near the contact of the cumulate and tectonite zones in ophiolites. We have identified 1,124 individual podiform chromite deposits, based on a 100-meter spatial rule, and have compiled them in a database. Of these, 619 deposits have been used to create three new grade and tonnage models for podiform chromite deposits. The major podiform chromite model has a median tonnage of 11,000 metric tons and a mean grade of 45 percent Cr2O3. The minor podiform chromite model has a median tonnage of 100 metric tons and a mean grade of 43 percent Cr2O3. The banded podiform chromite model has a median tonnage of 650 metric tons and a mean grade of 42 percent Cr2O3. Observed frequency distributions are also given for grades of rhodium, iridium, ruthenium, palladium, and platinum. In resource assessment applications, both major and minor podiform chromite models may be used for any ophiolite complex regardless of its tectonic setting or ophiolite zone. Expected sizes of undiscovered podiform chromite deposits, with respect to degree of deformation or ore-forming process, may determine which model is appropriate. The banded podiform chromite model may be applicable for ophiolites in both suprasubduction and midoceanic ridge settings.

  14. dSED: A database tool for modeling sediment early diagenesis

    NASA Astrophysics Data System (ADS)

    Katsev, S.; Rancourt, D. G.; L'Heureux, I.

    2003-04-01

    Sediment early diagenesis reaction transport models (RTMs) are becoming powerful tools in providing kinetic descriptions of the metal and nutrient diagenetic cycling in marine, lacustrine, estuarine, and other aquatic sediments, as well as of exchanges with the water column. Whereas there exist several good database/program combinations for thermodynamic equilibrium calculations in aqueous systems, at present there exist no database tools for classification and analysis of the kinetic data essential to RTM development. We present a database tool that is intended to serve as an online resource for information about chemical reactions, solid phase and solute reactants, sorption reactions, transport mechanisms, and kinetic and equilibrium parameters that are relevant to sediment diagenesis processes. The list of reactive substances includes but is not limited to organic matter, Fe and Mn oxides and oxyhydroxides, sulfides and sulfates, calcium, iron, and manganese carbonates, phosphorus-bearing minerals, and silicates. Aqueous phases include dissolved carbon dioxide, oxygen, methane, hydrogen sulfide, sulfate, nitrate, phosphate, some organic compounds, and dissolved metal species. A number of filters allow extracting information according to user-specified criteria, e.g., about a class of substances contributing to the cycling of iron. The database also includes bibliographic information about published diagenetic models and the reactions and processes that they consider. At the time of preparing this abstract, dSED contained 128 reactions and 12 pre-defined filters. dSED is maintained by the Lake Sediment Structure and Evolution (LSSE) group at the University of Ottawa (www.science.uottawa.ca/LSSE/dSED) and we invite input from the geochemical community.

  15. Generic models of deep formation water calculated with PHREEQC using the "gebo"-database

    NASA Astrophysics Data System (ADS)

    Bozau, E.; van Berk, W.

    2012-04-01

    To identify processes during the use of formation waters for geothermal energy production an extended hydrogeochemical thermodynamic database (named "gebo"-database) for the well known and commonly used software PHREEQC has been developed by collecting and inserting data from literature. The following solution master species: Fe(+2), Fe(+3), S(-2), C(-4), Si, Zn, Pb, and Al are added to the database "pitzer.dat" which is provided with the code PHREEQC. According to the solution master species the necessary solution species and phases (solid phases and gases) are implemented. Furthermore, temperature and pressure adaptations of the mass action law constants, Pitzer parameters for the calculation of activity coefficients in waters of high ionic strength and solubility equilibria among gaseous and aqueous species of CO2, methane, and hydrogen sulphide are implemented into the "gebo"-database. Combined with the "gebo"-database the code PHREEQC can be used to test the behaviour of highly concentrated solutions (e.g. formation waters, brines). Chemical changes caused by temperature and pressure gradients as well as the exposure of the water to the atmosphere and technical equipments can be modelled. To check the plausibility of additional and adapted data/parameters experimental solubility data from literature (e.g. sulfate and carbonate minerals) are compared to modelled mineral solubilities at elevated levels of Total Dissolved Solids (TDS), temperature, and pressure. First results show good matches between modelled and experimental mineral solubility for barite, celestite, anhydrite, and calcite in high TDS waters indicating the plausibility of additional and adapted data and parameters. Furthermore, chemical parameters of geothermal wells in the North German Basin are used to test the "gebo"-database. The analysed water composition (starting with the main cations and anions) is calculated by thermodynamic equilibrium reactions of pure water with the minerals found in

  16. Determination of urban volatile organic compound emission ratios and comparison with an emissions database

    NASA Astrophysics Data System (ADS)

    Warneke, C.; McKeen, S. A.; de Gouw, J. A.; Goldan, P. D.; Kuster, W. C.; Holloway, J. S.; Williams, E. J.; Lerner, B. M.; Parrish, D. D.; Trainer, M.; Fehsenfeld, F. C.; Kato, S.; Atlas, E. L.; Baker, A.; Blake, D. R.

    2007-05-01

    During the NEAQS-ITCT2k4 campaign in New England, anthropogenic VOCs and CO were measured downwind from New York City and Boston. The emission ratios of VOCs relative to CO and acetylene were calculated using a method in which the ratio of a VOC with acetylene is plotted versus the photochemical age. The intercept at the photochemical age of zero gives the emission ratio. The so determined emission ratios were compared to other measurement sets, including data from the same location in 2002, canister samples collected inside New York City and Boston, aircraft measurements from Los Angeles in 2002, and the average urban composition of 39 U.S. cities. All the measurements generally agree within a factor of two. The measured emission ratios also agree for most compounds within a factor of two with vehicle exhaust data indicating that a major source of VOCs in urban areas is automobiles. A comparison with an anthropogenic emission database shows less agreement. Especially large discrepancies were found for the C2-C4 alkanes and most oxygenated species. As an example, the database overestimated toluene by almost a factor of three, which caused an air quality forecast model (WRF-CHEM) using this database to overpredict the toluene mixing ratio by a factor of 2.5 as well. On the other hand, the overall reactivity of the measured species and the reactivity of the same compounds in the emission database were found to agree within 30%.

  17. Global and Regional Ecosystem Modeling: Databases of Model Drivers and Validation Measurements

    SciTech Connect

    Olson, R.J.

    2002-03-19

    }-grid cells for which inventory, modeling, or remote-sensing tools were used to scale up the point measurements. Documentation of the content and organization of the EMDI databases are provided.

  18. Data model and relational database design for the New England Water-Use Data System (NEWUDS)

    USGS Publications Warehouse

    Tessler, Steven

    2001-01-01

    The New England Water-Use Data System (NEWUDS) is a database for the storage and retrieval of water-use data. NEWUDS can handle data covering many facets of water use, including (1) tracking various types of water-use activities (withdrawals, returns, transfers, distributions, consumptive-use, wastewater collection, and treatment); (2) the description, classification and location of places and organizations involved in water-use activities; (3) details about measured or estimated volumes of water associated with water-use activities; and (4) information about data sources and water resources associated with water use. In NEWUDS, each water transaction occurs unidirectionally between two site objects, and the sites and conveyances form a water network. The core entities in the NEWUDS model are site, conveyance, transaction/rate, location, and owner. Other important entities include water resources (used for withdrawals and returns), data sources, and aliases. Multiple water-exchange estimates can be stored for individual transactions based on different methods or data sources. Storage of user-defined details is accommodated for several of the main entities. Numerous tables containing classification terms facilitate detailed descriptions of data items and can be used for routine or custom data summarization. NEWUDS handles single-user and aggregate-user water-use data, can be used for large or small water-network projects, and is available as a stand-alone Microsoft? Access database structure. Users can customize and extend the database, link it to other databases, or implement the design in other relational database applications.

  19. Geospatial Database for Strata Objects Based on Land Administration Domain Model (ladm)

    NASA Astrophysics Data System (ADS)

    Nasorudin, N. N.; Hassan, M. I.; Zulkifli, N. A.; Rahman, A. Abdul

    2016-09-01

    Recently in our country, the construction of buildings become more complex and it seems that strata objects database becomes more important in registering the real world as people now own and use multilevel of spaces. Furthermore, strata title was increasingly important and need to be well-managed. LADM is a standard model for land administration and it allows integrated 2D and 3D representation of spatial units. LADM also known as ISO 19152. The aim of this paper is to develop a strata objects database using LADM. This paper discusses the current 2D geospatial database and needs for 3D geospatial database in future. This paper also attempts to develop a strata objects database using a standard data model (LADM) and to analyze the developed strata objects database using LADM data model. The current cadastre system in Malaysia includes the strata title is discussed in this paper. The problems in the 2D geospatial database were listed and the needs for 3D geospatial database in future also is discussed. The processes to design a strata objects database are conceptual, logical and physical database design. The strata objects database will allow us to find the information on both non-spatial and spatial strata title information thus shows the location of the strata unit. This development of strata objects database may help to handle the strata title and information.

  20. Spatial Models for Architectural Heritage in Urban Database Context

    NASA Astrophysics Data System (ADS)

    Costamagna, E.; Spanò, A.

    2011-08-01

    Despite the GIS (Geographic Information Systems/Geospatial Information Systems) have been provided with several applications to manage the two-dimensional geometric information and arrange the topological relations among different spatial primitives, most of these systems have limited capabilities to manage the three-dimensional space. Other tools, such as CAD systems, have already achieved a full capability of representing 3D data. Most of the researches in the field of GIS have underlined the necessity of a full 3D management capability which is not yet achieved by the available systems (Rahman, Pilouk 2008) (Zlatanova 2002). First of all to reach this goal is important to define the spatial data model, which is at the same time a geometric and topological model and so integrating these two aspects in relation to the database management efficiency and documentation purposes. The application field on which these model can be tested is the spatial data managing of Architectural Heritage documentation, to evaluate the pertinence of these spatial models to the requested scale for the needs of such a documentation. Most of the important aspects are the integration of metric data originated from different sources and the representation and management of multiscale data. The issues connected with the representation of objects at higher LOD than the ones defined by the CityGML will be taken into account. The aim of this paper is then to investigate which are the favorable application of a framework in order to integrate two different approaches: architectural heritage spatial documentation and urban scale spatial data management.

  1. GIS-based hydrogeological databases and groundwater modelling

    NASA Astrophysics Data System (ADS)

    Gogu, Radu Constantin; Carabin, Guy; Hallet, Vincent; Peters, Valerie; Dassargues, Alain

    2001-12-01

    Reliability and validity of groundwater analysis strongly depend on the availability of large volumes of high-quality data. Putting all data into a coherent and logical structure supported by a computing environment helps ensure validity and availability and provides a powerful tool for hydrogeological studies. A hydrogeological geographic information system (GIS) database that offers facilities for groundwater-vulnerability analysis and hydrogeological modelling has been designed in Belgium for the Walloon region. Data from five river basins, chosen for their contrasting hydrogeological characteristics, have been included in the database, and a set of applications that have been developed now allow further advances. Interest is growing in the potential for integrating GIS technology and groundwater simulation models. A "loose-coupling" tool was created between the spatial-database scheme and the groundwater numerical model interface GMS (Groundwater Modelling System). Following time and spatial queries, the hydrogeological data stored in the database can be easily used within different groundwater numerical models. Résumé. La validité et la reproductibilité de l'analyse d'un aquifère dépend étroitement de la disponibilité de grandes quantités de données de très bonne qualité. Le fait de mettre toutes les données dans une structure cohérente et logique soutenue par les logiciels nécessaires aide à assurer la validité et la disponibilité et fournit un outil puissant pour les études hydrogéologiques. Une base de données pour un système d'information géographique (SIG) hydrogéologique qui offre toutes les facilités pour l'analyse de la vulnérabilité des eaux souterraines et la modélisation hydrogéologique a été établi en Belgique pour la région Wallonne. Les données de cinq bassins de rivières, choisis pour leurs caractéristiques hydrogéologiques différentes, ont été introduites dans la base de données, et un ensemble d

  2. Data-based modelling of the Earth's dynamic magnetosphere: a review

    NASA Astrophysics Data System (ADS)

    Tsyganenko, N. A.

    2013-10-01

    This paper reviews the main advances in the area of data-based modelling of the Earth's distant magnetic field achieved during the last two decades. The essence and the principal goal of the approach is to extract maximum information from available data, using physically realistic and flexible mathematical structures, parameterized by the most relevant and routinely accessible observables. Accordingly, the paper concentrates on three aspects of the modelling: (i) mathematical methods to develop a computational "skeleton" of a model, (ii) spacecraft databases, and (iii) parameterization of the magnetospheric models by the solar wind drivers and/or ground-based indices. The review is followed by a discussion of the main issues concerning further progress in the area, in particular, methods to assess the models' performance and the accuracy of the field line mapping. The material presented in the paper is organized along the lines of the author Julius-Bartels' Medal Lecture during the General Assembly 2013 of the European Geosciences Union.

  3. Virtual Organizations: Trends and Models

    NASA Astrophysics Data System (ADS)

    Nami, Mohammad Reza; Malekpour, Abbaas

    The Use of ICT in business has changed views about traditional business. With VO, organizations with out physical, geographical, or structural constraint can collaborate with together in order to fulfill customer requests in a networked environment. This idea improves resource utilization, reduces development process and costs, and saves time. Virtual Organization (VO) is always a form of partnership and managing partners and handling partnerships are crucial. Virtual organizations are defined as a temporary collection of enterprises that cooperate and share resources, knowledge, and competencies to better respond to business opportunities. This paper presents an overview of virtual organizations and main issues in collaboration such as security and management. It also presents a number of different model approaches according to their purpose and applications.

  4. Discovery of Possible Gene Relationships through the Application of Self-Organizing Maps to DNA Microarray Databases

    PubMed Central

    Chavez-Alvarez, Rocio; Chavoya, Arturo; Mendez-Vazquez, Andres

    2014-01-01

    DNA microarrays and cell cycle synchronization experiments have made possible the study of the mechanisms of cell cycle regulation of Saccharomyces cerevisiae by simultaneously monitoring the expression levels of thousands of genes at specific time points. On the other hand, pattern recognition techniques can contribute to the analysis of such massive measurements, providing a model of gene expression level evolution through the cell cycle process. In this paper, we propose the use of one of such techniques –an unsupervised artificial neural network called a Self-Organizing Map (SOM)–which has been successfully applied to processes involving very noisy signals, classifying and organizing them, and assisting in the discovery of behavior patterns without requiring prior knowledge about the process under analysis. As a test bed for the use of SOMs in finding possible relationships among genes and their possible contribution in some biological processes, we selected 282 S. cerevisiae genes that have been shown through biological experiments to have an activity during the cell cycle. The expression level of these genes was analyzed in five of the most cited time series DNA microarray databases used in the study of the cell cycle of this organism. With the use of SOM, it was possible to find clusters of genes with similar behavior in the five databases along two cell cycles. This result suggested that some of these genes might be biologically related or might have a regulatory relationship, as was corroborated by comparing some of the clusters obtained with SOMs against a previously reported regulatory network that was generated using biological knowledge, such as protein-protein interactions, gene expression levels, metabolism dynamics, promoter binding, and modification, regulation and transport of proteins. The methodology described in this paper could be applied to the study of gene relationships of other biological processes in different organisms. PMID:24699245

  5. The Mouse Genome Database (MGD): mouse biology and model systems.

    PubMed

    Bult, Carol J; Eppig, Janan T; Kadin, James A; Richardson, Joel E; Blake, Judith A

    2008-01-01

    The Mouse Genome Database, (MGD, http://www.informatics.jax.org/), integrates genetic, genomic and phenotypic information about the laboratory mouse, a primary animal model for studying human biology and disease. MGD data content includes comprehensive characterization of genes and their functions, standardized descriptions of mouse phenotypes, extensive integration of DNA and protein sequence data, normalized representation of genome and genome variant information including comparative data on mammalian genes. Data within MGD are obtained from diverse sources including manual curation of the biomedical literature, direct contributions from individual investigator's laboratories and major informatics resource centers such as Ensembl, UniProt and NCBI. MGD collaborates with the bioinformatics community on the development of data and semantic standards such as the Gene Ontology (GO) and the Mammalian Phenotype (MP) Ontology. MGD provides a data-mining platform that enables the development of translational research hypotheses based on comparative genotype, phenotype and functional analyses. Both web-based querying and computational access to data are provided. Recent improvements in MGD described here include the association of gene trap data with mouse genes and a new batch query capability for customized data access and retrieval.

  6. Filling a missing link between biogeochemical, climate and ecosystem studies: a global database of atmospheric water-soluble organic nitrogen

    NASA Astrophysics Data System (ADS)

    Cornell, Sarah

    2015-04-01

    It is time to collate a global community database of atmospheric water-soluble organic nitrogen deposition. Organic nitrogen (ON) has long been known to be globally ubiquitous in atmospheric aerosol and precipitation, with implications for air and water quality, climate, biogeochemical cycles, ecosystems and human health. The number of studies of atmospheric ON deposition has increased steadily in recent years, but to date there is no accessible global dataset, for either bulk ON or its major components. Improved qualitative and quantitative understanding of the organic nitrogen component is needed to complement the well-established knowledge base pertaining to other components of atmospheric deposition (cf. Vet et al 2014). Without this basic information, we are increasingly constrained in addressing the current dynamics and potential interactions of atmospheric chemistry, climate and ecosystem change. To see the full picture we need global data synthesis, more targeted data gathering, and models that let us explore questions about the natural and anthropogenic dynamics of atmospheric ON. Collectively, our research community already has a substantial amount of atmospheric ON data. Published reports extend back over a century and now have near-global coverage. However, datasets available from the literature are very piecemeal and too often lack crucially important information that would enable aggregation or re-use. I am initiating an open collaborative process to construct a community database, so we can begin to systematically synthesize these datasets (generally from individual studies at a local and temporally limited scale) to increase their scientific usability and statistical power for studies of global change and anthropogenic perturbation. In drawing together our disparate knowledge, we must address various challenges and concerns, not least about the comparability of analysis and sampling methodologies, and the known complexity of composition of ON. We

  7. A database for estimating organ dose for coronary angiography and brain perfusion CT scans for arbitrary spectra and angular tube current modulation

    SciTech Connect

    Rupcich, Franco; Badal, Andreu; Kyprianou, Iacovos; Schmidt, Taly Gilat

    2012-09-15

    Purpose: The purpose of this study was to develop a database for estimating organ dose in a voxelized patient model for coronary angiography and brain perfusion CT acquisitions with any spectra and angular tube current modulation setting. The database enables organ dose estimation for existing and novel acquisition techniques without requiring Monte Carlo simulations. Methods: The study simulated transport of monoenergetic photons between 5 and 150 keV for 1000 projections over 360 Degree-Sign through anthropomorphic voxelized female chest and head (0 Degree-Sign and 30 Degree-Sign tilt) phantoms and standard head and body CTDI dosimetry cylinders. The simulations resulted in tables of normalized dose deposition for several radiosensitive organs quantifying the organ dose per emitted photon for each incident photon energy and projection angle for coronary angiography and brain perfusion acquisitions. The values in a table can be multiplied by an incident spectrum and number of photons at each projection angle and then summed across all energies and angles to estimate total organ dose. Scanner-specific organ dose may be approximated by normalizing the database-estimated organ dose by the database-estimated CTDI{sub vol} and multiplying by a physical CTDI{sub vol} measurement. Two examples are provided demonstrating how to use the tables to estimate relative organ dose. In the first, the change in breast and lung dose during coronary angiography CT scans is calculated for reduced kVp, angular tube current modulation, and partial angle scanning protocols relative to a reference protocol. In the second example, the change in dose to the eye lens is calculated for a brain perfusion CT acquisition in which the gantry is tilted 30 Degree-Sign relative to a nontilted scan. Results: Our database provides tables of normalized dose deposition for several radiosensitive organs irradiated during coronary angiography and brain perfusion CT scans. Validation results indicate

  8. Java Web Simulation (JWS); a web based database of kinetic models.

    PubMed

    Snoep, J L; Olivier, B G

    2002-01-01

    Software to make a database of kinetic models accessible via the internet has been developed and a core database has been set up at http://jjj.biochem.sun.ac.za/. This repository of models, available to everyone with internet access, opens a whole new way in which we can make our models public. Via the database, a user can change enzyme parameters and run time simulations or steady state analyses. The interface is user friendly and no additional software is necessary. The database currently contains 10 models, but since the generation of the program code to include new models has largely been automated the addition of new models is straightforward and people are invited to submit their models to be included in the database.

  9. Modeling of heavy organic deposition

    SciTech Connect

    Chung, F.T.H.

    1992-01-01

    Organic deposition is often a major problem in petroleum production and processing. This problem is manifested by current activities in gas flooding and heavy oil production. The need for understanding the nature of asphaltenes and asphaltics and developing solutions to the deposition problem is well recognized. Prediction technique is crucial to solution development. In the past 5 years, some progress in modeling organic deposition has been made. A state-of-the-art review of methods for modeling organic deposition is presented in this report. Two new models were developed in this work; one based on a thermodynamic equilibrium principle and the other on the colloidal stability theory. These two models are more general and realistic than others previously reported. Because experimental results on the characteristics of asphaltene are inconclusive, it is still not well known whether the asphaltenes is crude oil exist as a true solution or as a colloidal suspension. Further laboratory work which is designed to study the solubility properties of asphaltenes and to provide additional information for model development is proposed. Some experimental tests have been conducted to study the mechanisms of CO{sub 2}-induced asphaltene precipitation. Coreflooding experiments show that asphaltene precipitation occurs after gas breakthrough. The mechanism of CO{sub 2}-induced asphaltene precipitation is believed to occur by hydrocarbon extraction which causes change in oil composition. Oil swelling due to CO{sub 2} solubilization does not induce asphaltene precipitation.

  10. Models for financial sustainability of biological databases and resources.

    PubMed

    Chandras, Christina; Weaver, Thomas; Zouberakis, Michael; Smedley, Damian; Schughart, Klaus; Rosenthal, Nadia; Hancock, John M; Kollias, George; Schofield, Paul N; Aidinis, Vassilis

    2009-01-01

    Following the technological advances that have enabled genome-wide analysis in most model organisms over the last decade, there has been unprecedented growth in genomic and post-genomic science with concomitant generation of an exponentially increasing volume of data and material resources. As a result, numerous repositories have been created to store and archive data, organisms and material, which are of substantial value to the whole community. Sustained access, facilitating re-use of these resources, is essential, not only for validation, but for re-analysis, testing of new hypotheses and developing new technologies/platforms. A common challenge for most data resources and biological repositories today is finding financial support for maintenance and development to best serve the scientific community. In this study we examine the problems that currently confront the data and resource infrastructure underlying the biomedical sciences. We discuss the financial sustainability issues and potential business models that could be adopted by biological resources and consider long term preservation issues within the context of mouse functional genomics efforts in Europe.

  11. Database Administrator

    ERIC Educational Resources Information Center

    Moore, Pam

    2010-01-01

    The Internet and electronic commerce (e-commerce) generate lots of data. Data must be stored, organized, and managed. Database administrators, or DBAs, work with database software to find ways to do this. They identify user needs, set up computer databases, and test systems. They ensure that systems perform as they should and add people to the…

  12. Avibase – a database system for managing and organizing taxonomic concepts

    PubMed Central

    Lepage, Denis; Vaidya, Gaurav; Guralnick, Robert

    2014-01-01

    Abstract Scientific names of biological entities offer an imperfect resolution of the concepts that they are intended to represent. Often they are labels applied to entities ranging from entire populations to individual specimens representing those populations, even though such names only unambiguously identify the type specimen to which they were originally attached. Thus the real-life referents of names are constantly changing as biological circumscriptions are redefined and thereby alter the sets of individuals bearing those names. This problem is compounded by other characteristics of names that make them ambiguous identifiers of biological concepts, including emendations, homonymy and synonymy. Taxonomic concepts have been proposed as a way to address issues related to scientific names, but they have yet to receive broad recognition or implementation. Some efforts have been made towards building systems that address these issues by cataloguing and organizing taxonomic concepts, but most are still in conceptual or proof-of-concept stage. We present the on-line database Avibase as one possible approach to organizing taxonomic concepts. Avibase has been successfully used to describe and organize 844,000 species-level and 705,000 subspecies-level taxonomic concepts across every major bird taxonomic checklist of the last 125 years. The use of taxonomic concepts in place of scientific names, coupled with efficient resolution services, is a major step toward addressing some of the main deficiencies in the current practices of scientific name dissemination and use. PMID:25061375

  13. FOAM (Functional Ontology Assignments for Metagenomes): A Hidden Markov Model (HMM) database with environmental focus

    SciTech Connect

    Prestat, Emmanuel; David, Maude M.; Hultman, Jenni; Ta , Neslihan; Lamendella, Regina; Dvornik, Jill; Mackelprang, Rachel; Myrold, David D.; Jumpponen, Ari; Tringe, Susannah G.; Holman, Elizabeth; Mavromatis, Konstantinos; Jansson, Janet K.

    2014-09-26

    A new functional gene database, FOAM (Functional Ontology Assignments for Metagenomes), was developed to screen environmental metagenomic sequence datasets. FOAM provides a new functional ontology dedicated to classify gene functions relevant to environmental microorganisms based on Hidden Markov Models (HMMs). Sets of aligned protein sequences (i.e. ‘profiles’) were tailored to a large group of target KEGG Orthologs (KOs) from which HMMs were trained. The alignments were checked and curated to make them specific to the targeted KO. Within this process, sequence profiles were enriched with the most abundant sequences available to maximize the yield of accurate classifier models. An associated functional ontology was built to describe the functional groups and hierarchy. FOAM allows the user to select the target search space before HMM-based comparison steps and to easily organize the results into different functional categories and subcategories. FOAM is publicly available at http://portal.nersc.gov/project/m1317/FOAM/.

  14. FOAM (Functional Ontology Assignments for Metagenomes): a Hidden Markov Model (HMM) database with environmental focus.

    PubMed

    Prestat, Emmanuel; David, Maude M; Hultman, Jenni; Taş, Neslihan; Lamendella, Regina; Dvornik, Jill; Mackelprang, Rachel; Myrold, David D; Jumpponen, Ari; Tringe, Susannah G; Holman, Elizabeth; Mavromatis, Konstantinos; Jansson, Janet K

    2014-10-29

    A new functional gene database, FOAM (Functional Ontology Assignments for Metagenomes), was developed to screen environmental metagenomic sequence datasets. FOAM provides a new functional ontology dedicated to classify gene functions relevant to environmental microorganisms based on Hidden Markov Models (HMMs). Sets of aligned protein sequences (i.e. 'profiles') were tailored to a large group of target KEGG Orthologs (KOs) from which HMMs were trained. The alignments were checked and curated to make them specific to the targeted KO. Within this process, sequence profiles were enriched with the most abundant sequences available to maximize the yield of accurate classifier models. An associated functional ontology was built to describe the functional groups and hierarchy. FOAM allows the user to select the target search space before HMM-based comparison steps and to easily organize the results into different functional categories and subcategories. FOAM is publicly available at http://portal.nersc.gov/project/m1317/FOAM/.

  15. Solid Waste Projection Model: Database (Version 1.3). Technical reference manual

    SciTech Connect

    Blackburn, C.L.

    1991-11-01

    The Solid Waste Projection Model (SWPM) system is an analytical tool developed by Pacific Northwest Laboratory (PNL) for Westinghouse Hanford Company (WHC). The SWPM system provides a modeling and analysis environment that supports decisions in the process of evaluating various solid waste management alternatives. This document, one of a series describing the SWPM system, contains detailed information regarding the software and data structures utilized in developing the SWPM Version 1.3 Database. This document is intended for use by experienced database specialists and supports database maintenance, utility development, and database enhancement.

  16. Historical Land Use Change Estimates for Climate Modelers: Results from The HYDE Database.

    NASA Astrophysics Data System (ADS)

    Klein Goldewijk, K.

    2003-04-01

    It is beyond doubt that human activities always have modified the natural environment, but it has become clear that during the last centuries the intensity and scale of these modifications has increased dramatically. Land cover changes affect climate by their impact on surface energy and moisture budgets, and thus should be included in global climate models. Therefore, a growing need is developed for better knowledge of historical land cover. A database with historical data of the global environment (HYDE) was created, which can be used in global climate models. HYDE covers not only land use (changes), but also general topics such as population, livestock, gross domestic product, and value added of industry and/or services as well. These driving forces occur at several spatial and temporal scales and dimensions, and differ often among regions. This requires a geographically explicit modeling approach. Where possible, data have been organized at the country level, and for the period 1700 to 1990. Some data are also available with geographic detail (Klein Goldewijk, 2001; Klein Goldewijk and Battjes, 1997). Examples of a global reconstruction of 300 years historical land use are presented, using gridded historical population estimates as a proxy for allocation of agricultural land. References: Klein Goldewijk, K., 2001. Estimating Global Land Use over the past 300 years: The HYDE 2.0 database. Global Biogeochemical Cycles 15(2): 417--433. Klein Goldewijk, C.G.M. and J.J. Battjes, 1997. A Hundred Year (1890 1990) Database for Integrated Environmental Assessments (HYDE, version 1.1). RIVM Report no. 422514002. National Institute of Public Health and Environmental Protection (RIVM). 196 pp. Internet: http://www.rivm.nl/env/int/hyde/

  17. The model organism Dictyostelium discoideum.

    PubMed

    Bozzaro, Salvatore

    2013-01-01

    Much of our knowledge of molecular cellular functions is based on studies with a few number of model organisms that were established during the last 50 years. The social amoeba Dictyostelium discoideum is one such model, and has been particularly useful for the study of cell motility, chemotaxis, phagocytosis, endocytic vesicle traffic, cell adhesion, pattern formation, caspase-independent cell death, and, more recently, autophagy and social evolution. As nonmammalian model of human diseases D. discoideum is a newcomer, yet it has proven to be a powerful genetic and cellular model for investigating host-pathogen interactions and microbial infections, for mitochondrial diseases, and for pharmacogenetic studies. The D. discoideum genome harbors several homologs of human genes responsible for a variety of diseases, -including Chediak-Higashi syndrome, lissencephaly, mucolipidosis, Huntington disease, IBMPFD, and Shwachman-Diamond syndrome. A few genes have already been studied, providing new insights on the mechanism of action of the encoded proteins and in some cases on the defect underlying the disease. The opportunities offered by the organism and its place among the nonmammalian models for human diseases will be discussed. PMID:23494300

  18. Solid Waste Projection Model: Database (Version 1.4). Technical reference manual

    SciTech Connect

    Blackburn, C.; Cillan, T.

    1993-09-01

    The Solid Waste Projection Model (SWPM) system is an analytical tool developed by Pacific Northwest Laboratory (PNL) for Westinghouse Hanford Company (WHC). The SWPM system provides a modeling and analysis environment that supports decisions in the process of evaluating various solid waste management alternatives. This document, one of a series describing the SWPM system, contains detailed information regarding the software and data structures utilized in developing the SWPM Version 1.4 Database. This document is intended for use by experienced database specialists and supports database maintenance, utility development, and database enhancement. Those interested in using the SWPM database should refer to the SWPM Database User`s Guide. This document is available from the PNL Task M Project Manager (D. L. Stiles, 509-372-4358), the PNL Task L Project Manager (L. L. Armacost, 509-372-4304), the WHC Restoration Projects Section Manager (509-372-1443), or the WHC Waste Characterization Manager (509-372-1193).

  19. Database specification for the Worldwide Port System (WPS) Regional Integrated Cargo Database (ICDB)

    SciTech Connect

    Faby, E.Z.; Fluker, J.; Hancock, B.R.; Grubb, J.W.; Russell, D.L.; Loftis, J.P.; Shipe, P.C.; Truett, L.F.

    1994-03-01

    This Database Specification for the Worldwide Port System (WPS) Regional Integrated Cargo Database (ICDB) describes the database organization and storage allocation, provides the detailed data model of the logical and physical designs, and provides information for the construction of parts of the database such as tables, data elements, and associated dictionaries and diagrams.

  20. Modeling BVOC isoprene emissions based on a GIS and remote sensing database

    NASA Astrophysics Data System (ADS)

    Wong, Man Sing; Sarker, Md. Latifur Rahman; Nichol, Janet; Lee, Shun-cheng; Chen, Hongwei; Wan, Yiliang; Chan, P. W.

    2013-04-01

    This paper presents a geographic information systems (GIS) model to relate biogenic volatile organic compounds (BVOCs) isoprene emissions to ecosystem type, as well as environmental drivers such as light intensity, temperature, landscape factor and foliar density. Data and techniques have recently become available which can permit new improved estimates of isoprene emissions over Hong Kong. The techniques are based on Guenther et al.'s (1993, 1999) model. The spatially detailed mapping of isoprene emissions over Hong Kong at a resolution of 100 m and a database has been constructed for retrieval of the isoprene maps from February 2007 to January 2008. This approach assigns emission rates directly to ecosystem types not to individual species, since unlike in temperate regions where one or two single species may dominate over large regions, Hong Kong's vegetation is extremely diverse with up to 300 different species in 1 ha. Field measurements of emissions by canister sampling obtained a range of ambient emissions according to different climatic conditions for Hong Kong's main ecosystem types in both urban and rural areas, and these were used for model validation. Results show the model-derived isoprene flux to have high to moderate correlations with field observations (i.e. r2 = 0.77, r2 = 0.63, r2 = 0.37 for all 24 field measurements, subset for summer, and winter data, respectively) which indicate the robustness of the approach when applied to tropical forests at detailed level, as well as the promising role of remote sensing in isoprene mapping. The GIS model and raster database provide a simple and low cost estimation of the BVOC isoprene in Hong Kong at detailed level. City planners and environmental authorities may use the derived models for estimating isoprene transportation, and its interaction with anthropogenic pollutants in urban areas.

  1. Derivation of rules for comparative protein modeling from a database of protein structure alignments.

    PubMed Central

    Sali, A.; Overington, J. P.

    1994-01-01

    We describe a database of protein structure alignments as well as methods and tools that use this database to improve comparative protein modeling. The current version of the database contains 105 alignments of similar proteins or protein segments. The database comprises 416 entries, 78,495 residues, 1,233 equivalent entry pairs, and 230,396 pairs of equivalent alignment positions. At present, the main application of the database is to improve comparative modeling by satisfaction of spatial restraints implemented in the program MODELLER (Sali A, Blundell TL, 1993, J Mol Biol 234:779-815). To illustrate the usefulness of the database, the restraints on the conformation of a disulfide bridge provided by an equivalent disulfide bridge in a related structure are derived from the alignments; the prediction success of the disulfide dihedral angle classes is increased to approximately 80%, compared to approximately 55% for modeling that relies on the stereochemistry of disulfide bridges alone. The second example of the use of the database is the derivation of the probability density function for comparative modeling of the cis/trans isomerism of the proline residues; the prediction success is increased from 0% to 82.9% for cis-proline and from 93.3% to 96.2% for trans-proline. The database is available via electronic mail. PMID:7833817

  2. Functional Analysis and Discovery of Microbial Genes Transforming Metallic and Organic Pollutants: Database and Experimental Tools

    SciTech Connect

    Lawrence P. Wackett; Lynda B.M. Ellis

    2004-12-09

    Microbial functional genomics is faced with a burgeoning list of genes which are denoted as unknown or hypothetical for lack of any knowledge about their function. The majority of microbial genes encode enzymes. Enzymes are the catalysts of metabolism; catabolism, anabolism, stress responses, and many other cell functions. A major problem facing microbial functional genomics is proposed here to derive from the breadth of microbial metabolism, much of which remains undiscovered. The breadth of microbial metabolism has been surveyed by the PIs and represented according to reaction types on the University of Minnesota Biocatalysis/Biodegradation Database (UM-BBD): http://umbbd.ahc.umn.edu/search/FuncGrps.html The database depicts metabolism of 49 chemical functional groups, representing most of current knowledge. Twice that number of chemical groups are proposed here to be metabolized by microbes. Thus, at least 50% of the unique biochemical reactions catalyzed by microbes remain undiscovered. This further suggests that many unknown and hypothetical genes encode functions yet undiscovered. This gap will be partly filled by the current proposal. The UM-BBD will be greatly expanded as a resource for microbial functional genomics. Computational methods will be developed to predict microbial metabolism which is not yet discovered. Moreover, a concentrated effort to discover new microbial metabolism will be conducted. The research will focus on metabolism of direct interest to DOE, dealing with the transformation of metals, metalloids, organometallics and toxic organics. This is precisely the type of metabolism which has been characterized most poorly to date. Moreover, these studies will directly impact functional genomic analysis of DOE-relevant genomes.

  3. A Conceptual Modeling of spatio-temporal database to Estimate Runoff Changes in Urbanized Watersheds

    NASA Astrophysics Data System (ADS)

    Schäfer, A. L.

    2014-11-01

    This paper presents the development of a conceptual model of a database that allows the monitoring of changes in watersheds over time and verifies the impact of these changes on runoff. The conceptual model was developed using ER modeling techniques. ER diagrams were constructed from the semantic analysis of the variables involved in the issue of runoff in watersheds using the Curve Number method of Natural Resource Conservation Service. The conceptual model was developed based on the concepts of states and events, and the use of time as a basis for organizing spatial data allowed to record the time and place of any changes. The model of representation of spatial data based on object was employed. Through the proposed conceptual model, it is possible to store the results of several simulations for a watershed, linking each simulation to a specific event and identifying which scenario is valid at the time. Thus, it is possible to identify quantitative changes related to runoff over time and relate them to the events that caused them and the entities involved in such events. The conceptual model supports the existence of alternate realities, allowing the simulation and comparison of past and future scenarios.

  4. Integrated Functional and Executional Modelling of Software Using Web-Based Databases

    NASA Technical Reports Server (NTRS)

    Kulkarni, Deepak; Marietta, Roberta

    1998-01-01

    NASA's software subsystems undergo extensive modification and updates over the operational lifetimes. It is imperative that modified software should satisfy safety goals. This report discusses the difficulties encountered in doing so and discusses a solution based on integrated modelling of software, use of automatic information extraction tools, web technology and databases. To appear in an article of Journal of Database Management.

  5. DSSTOX WEBSITE LAUNCH: IMPROVING PUBLIC ACCESS TO DATABASES FOR BUILDING STRUCTURE-TOXICITY PREDICTION MODELS

    EPA Science Inventory

    DSSTox Website Launch: Improving Public Access to Databases for Building Structure-Toxicity Prediction Models
    Ann M. Richard
    US Environmental Protection Agency, Research Triangle Park, NC, USA

    Distributed: Decentralized set of standardized, field-delimited databases,...

  6. Teaching Database Modeling and Design: Areas of Confusion and Helpful Hints

    ERIC Educational Resources Information Center

    Philip, George C.

    2007-01-01

    This paper identifies several areas of database modeling and design that have been problematic for students and even are likely to confuse faculty. Major contributing factors are the lack of clarity and inaccuracies that persist in the presentation of some basic database concepts in textbooks. The paper analyzes the problems and discusses ways to…

  7. A Database for Propagation Models and Conversion to C++ Programming Language

    NASA Technical Reports Server (NTRS)

    Kantak, Anil V.; Angkasa, Krisjani; Rucker, James

    1996-01-01

    In the past few years, a computer program was produced to contain propagation models and the necessary prediction methods of most propagation phenomena. The propagation model database described here creates a user friendly environment that makes using the database easy for experienced users and novices alike. The database is designed to pass data through the desired models easily and generate relevant results quickly. The database already contains many of the propagation phenomena models accepted by the propagation community and every year new models are added. The major sources of models included are the NASA Propagation Effects Handbook or the International Radio Consultive Committee (CCIR) or publications such as the Institute for Electrical and Electronic Engineers (IEEE).

  8. Developing a comprehensive database management system for organization and evaluation of mammography datasets.

    PubMed

    Wu, Yirong; Rubin, Daniel L; Woods, Ryan W; Elezaby, Mai; Burnside, Elizabeth S

    2014-01-01

    We aimed to design and develop a comprehensive mammography database system (CMDB) to collect clinical datasets for outcome assessment and development of decision support tools. A Health Insurance Portability and Accountability Act (HIPAA) compliant CMDB was created to store multi-relational datasets of demographic risk factors and mammogram results using the Breast Imaging Reporting and Data System (BI-RADS) lexicon. The CMDB collected both biopsy pathology outcomes, in a breast pathology lexicon compiled by extending BI-RADS, and our institutional breast cancer registry. The audit results derived from the CMDB were in accordance with Mammography Quality Standards Act (MQSA) audits and national benchmarks. The CMDB has managed the challenges of multi-level organization demanded by the complexity of mammography practice and lexicon development in pathology. We foresee that the CMDB will be useful for efficient quality assurance audits and development of decision support tools to improve breast cancer diagnosis. Our procedure of developing the CMDB provides a framework to build a detailed data repository for breast imaging quality control and research, which has the potential to augment existing resources.

  9. Developing a Comprehensive Database Management System for Organization and Evaluation of Mammography Datasets

    PubMed Central

    Wu, Yirong; Rubin, Daniel L; Woods, Ryan W; Elezaby, Mai; Burnside, Elizabeth S

    2014-01-01

    We aimed to design and develop a comprehensive mammography database system (CMDB) to collect clinical datasets for outcome assessment and development of decision support tools. A Health Insurance Portability and Accountability Act (HIPAA) compliant CMDB was created to store multi-relational datasets of demographic risk factors and mammogram results using the Breast Imaging Reporting and Data System (BI-RADS) lexicon. The CMDB collected both biopsy pathology outcomes, in a breast pathology lexicon compiled by extending BI-RADS, and our institutional breast cancer registry. The audit results derived from the CMDB were in accordance with Mammography Quality Standards Act (MQSA) audits and national benchmarks. The CMDB has managed the challenges of multi-level organization demanded by the complexity of mammography practice and lexicon development in pathology. We foresee that the CMDB will be useful for efficient quality assurance audits and development of decision support tools to improve breast cancer diagnosis. Our procedure of developing the CMDB provides a framework to build a detailed data repository for breast imaging quality control and research, which has the potential to augment existing resources. PMID:25368510

  10. An online database for informing ecological network models: http://kelpforest.ucsc.edu

    USGS Publications Warehouse

    Beas-Luna, Rodrigo; Tinker, M. Tim; Novak, Mark; Carr, Mark H.; Black, August; Caselle, Jennifer E.; Hoban, Michael; Malone, Dan; Iles, Alison C.

    2014-01-01

    Ecological network models and analyses are recognized as valuable tools for understanding the dynamics and resiliency of ecosystems, and for informing ecosystem-based approaches to management. However, few databases exist that can provide the life history, demographic and species interaction information necessary to parameterize ecological network models. Faced with the difficulty of synthesizing the information required to construct models for kelp forest ecosystems along the West Coast of North America, we developed an online database (http://kelpforest.ucsc.edu/) to facilitate the collation and dissemination of such information. Many of the database's attributes are novel yet the structure is applicable and adaptable to other ecosystem modeling efforts. Information for each taxonomic unit includes stage-specific life history, demography, and body-size allometries. Species interactions include trophic, competitive, facilitative, and parasitic forms. Each data entry is temporally and spatially explicit. The online data entry interface allows researchers anywhere to contribute and access information. Quality control is facilitated by attributing each entry to unique contributor identities and source citations. The database has proven useful as an archive of species and ecosystem-specific information in the development of several ecological network models, for informing management actions, and for education purposes (e.g., undergraduate and graduate training). To facilitate adaptation of the database by other researches for other ecosystems, the code and technical details on how to customize this database and apply it to other ecosystems are freely available and located at the following link (https://github.com/kelpforest-cameo/data​baseui).

  11. An Online Database for Informing Ecological Network Models: http://kelpforest.ucsc.edu

    PubMed Central

    Beas-Luna, Rodrigo; Novak, Mark; Carr, Mark H.; Tinker, Martin T.; Black, August; Caselle, Jennifer E.; Hoban, Michael; Malone, Dan; Iles, Alison

    2014-01-01

    Ecological network models and analyses are recognized as valuable tools for understanding the dynamics and resiliency of ecosystems, and for informing ecosystem-based approaches to management. However, few databases exist that can provide the life history, demographic and species interaction information necessary to parameterize ecological network models. Faced with the difficulty of synthesizing the information required to construct models for kelp forest ecosystems along the West Coast of North America, we developed an online database (http://kelpforest.ucsc.edu/) to facilitate the collation and dissemination of such information. Many of the database's attributes are novel yet the structure is applicable and adaptable to other ecosystem modeling efforts. Information for each taxonomic unit includes stage-specific life history, demography, and body-size allometries. Species interactions include trophic, competitive, facilitative, and parasitic forms. Each data entry is temporally and spatially explicit. The online data entry interface allows researchers anywhere to contribute and access information. Quality control is facilitated by attributing each entry to unique contributor identities and source citations. The database has proven useful as an archive of species and ecosystem-specific information in the development of several ecological network models, for informing management actions, and for education purposes (e.g., undergraduate and graduate training). To facilitate adaptation of the database by other researches for other ecosystems, the code and technical details on how to customize this database and apply it to other ecosystems are freely available and located at the following link (https://github.com/kelpforest-cameo/databaseui). PMID:25343723

  12. Estimating spatial distribution of soil organic carbon for the Midwestern United States using historical database.

    PubMed

    Kumar, Sandeep

    2015-05-01

    Soil organic carbon (SOC) is the most important parameter influencing soil health, global climate change, crop productivity, and various ecosystem services. Therefore, estimating SOC at larger scales is important. The present study was conducted to estimate the SOC pool at regional scale using the historical database gathered by the National Soil Survey Staff. Specific objectives of the study were to upscale the SOC density (kg C m(-2)) and total SOC pool (PgC) across the Midwestern United States using the geographically weighted regression kriging (GWRK), and compare the results with those obtained from the geographically weighted regression (GWR) using the data for 3485 georeferenced profiles. Results from this study support the conclusion that the GWRK produced satisfactory predictions with lower root mean square error (5.60 kg m(-2)), mean estimation error (0.01 kg m(-2)) and mean absolute estimation error (4.30 kg m(-2)), and higher R(2) (0.58) and goodness-of-prediction statistic (G=0.59) values. The superiority of this approach is evident through a substantial increase in R(2) (0.45) compared to that for the global regression (R(2)=0.28). Croplands of the region store 16.8 Pg SOC followed by shrubs (5.85 Pg) and forests (4.45 Pg). Total SOC pool for the Midwestern region ranges from 31.5 to 31.6 Pg. This study illustrates that the GWRK approach explicitly addresses the spatial dependency and spatial non-stationarity issues for interpolating SOC density across the regional scale.

  13. Completion of autobuilt protein models using a database of protein fragments

    SciTech Connect

    Cowtan, Kevin

    2012-04-01

    Two developments in the process of automated protein model building in the Buccaneer software are described: the use of a database of protein fragments in improving the model completeness and the assembly of disconnected chain fragments into complete molecules. Two developments in the process of automated protein model building in the Buccaneer software are presented. A general-purpose library for protein fragments of arbitrary size is described, with a highly optimized search method allowing the use of a larger database than in previous work. The problem of assembling an autobuilt model into complete chains is discussed. This involves the assembly of disconnected chain fragments into complete molecules and the use of the database of protein fragments in improving the model completeness. Assembly of fragments into molecules is a standard step in existing model-building software, but the methods have not received detailed discussion in the literature.

  14. Database Manager

    ERIC Educational Resources Information Center

    Martin, Andrew

    2010-01-01

    It is normal practice today for organizations to store large quantities of records of related information as computer-based files or databases. Purposeful information is retrieved by performing queries on the data sets. The purpose of DATABASE MANAGER is to communicate to students the method by which the computer performs these queries. This…

  15. Guide on Data Models in the Selection and Use of Database Management Systems. Final Report.

    ERIC Educational Resources Information Center

    Gallagher, Leonard J.; Draper, Jesse M.

    A tutorial introduction to data models in general is provided, with particular emphasis on the relational and network models defined by the two proposed ANSI (American National Standards Institute) database language standards. Examples based on the network and relational models include specific syntax and semantics, while examples from the other…

  16. Automatic generation of conceptual database design tools from data model specifications

    SciTech Connect

    Hong, Shuguang.

    1989-01-01

    The problems faced in the design and implementation of database software systems based on object-oriented data models are similar to that of other software design, i.e., difficult, complex, yet redundant effort. Automatic generation of database software system has been proposed as a solution to the problems. In order to generate database software system for a variety of object-oriented data models, two critical issues: data model specification and software generation, must be addressed. SeaWeed is a software system that automatically generates conceptual database design tools from data model specifications. A meta model has been defined for the specification of a class of object-oriented data models. This meta model provides a set of primitive modeling constructs that can be used to express the semantics, or unique characteristics, of specific data models. Software reusability has been adopted for the software generation. The technique of design reuse is utilized to derive the requirement specification of the software to be generated from data model specifications. The mechanism of code reuse is used to produce the necessary reusable software components. This dissertation presents the research results of SeaWeed including the meta model, data model specification, a formal representation of design reuse and code reuse, and the software generation paradigm.

  17. Combining computational models, semantic annotations and simulation experiments in a graph database

    PubMed Central

    Henkel, Ron; Wolkenhauer, Olaf; Waltemath, Dagmar

    2015-01-01

    Model repositories such as the BioModels Database, the CellML Model Repository or JWS Online are frequently accessed to retrieve computational models of biological systems. However, their storage concepts support only restricted types of queries and not all data inside the repositories can be retrieved. In this article we present a storage concept that meets this challenge. It grounds on a graph database, reflects the models’ structure, incorporates semantic annotations and simulation descriptions and ultimately connects different types of model-related data. The connections between heterogeneous model-related data and bio-ontologies enable efficient search via biological facts and grant access to new model features. The introduced concept notably improves the access of computational models and associated simulations in a model repository. This has positive effects on tasks such as model search, retrieval, ranking, matching and filtering. Furthermore, our work for the first time enables CellML- and Systems Biology Markup Language-encoded models to be effectively maintained in one database. We show how these models can be linked via annotations and queried. Database URL: https://sems.uni-rostock.de/projects/masymos/ PMID:25754863

  18. Petri-net-based modeling and evaluation of pipelined processing of concurrent database queries

    SciTech Connect

    Mikkilimeni, K.P.; Chow, Y.C.; Su, S.Y.W.

    1988-11-01

    This paper describes a Petri-net-based methodology for modeling and evaluation pipelined processing of concurrent database queries in an integrated data network (IDN). An extended Petri-net model is presented to model two key approaches to concurrent data-base query processing in the IDN, namely, pipelined and data-flow-based execution of queries and intermediate data sharing among concurrent queries. Database operations are categorized, and the models for the data flow and control flow in them are presented. A general purpose Petri-net simulator has been developed using event-driven programming techniques. Using the simulator, the execution of the Petri-net models of some test queries has been simulated.

  19. Publication trends in model organism research.

    PubMed

    Dietrich, Michael R; Ankeny, Rachel A; Chen, Patrick M

    2014-11-01

    In 1990, the National Institutes of Health (NIH) gave some organisms special status as designated model organisms. This article documents publication trends for these NIH-designated model organisms over the past 40 years. We find that being designated a model organism by the NIH does not guarantee an increasing publication trend. An analysis of model and nonmodel organisms included in GENETICS since 1960 does reveal a sharp decline in the number of publications using nonmodel organisms yet no decline in the overall species diversity. We suggest that organisms with successful publication records tend to share critical characteristics, such as being well developed as standardized, experimental systems and being used by well-organized communities with good networks of exchange and methods of communication.

  20. Publication Trends in Model Organism Research

    PubMed Central

    Dietrich, Michael R.; Ankeny, Rachel A.; Chen, Patrick M.

    2014-01-01

    In 1990, the National Institutes of Health (NIH) gave some organisms special status as designated model organisms. This article documents publication trends for these NIH-designated model organisms over the past 40 years. We find that being designated a model organism by the NIH does not guarantee an increasing publication trend. An analysis of model and nonmodel organisms included in GENETICS since 1960 does reveal a sharp decline in the number of publications using nonmodel organisms yet no decline in the overall species diversity. We suggest that organisms with successful publication records tend to share critical characteristics, such as being well developed as standardized, experimental systems and being used by well-organized communities with good networks of exchange and methods of communication. PMID:25381363

  1. Analysis of the Properties of Working Substances for the Organic Rankine Cycle based Database "REFPROP"

    NASA Astrophysics Data System (ADS)

    Galashov, Nikolay; Tsibulskiy, Svyatoslav; Serova, Tatiana

    2016-02-01

    The object of the study are substances that are used as a working fluid in systems operating on the basis of an organic Rankine cycle. The purpose of research is to find substances with the best thermodynamic, thermal and environmental properties. Research conducted on the basis of the analysis of thermodynamic and thermal properties of substances from the base "REFPROP" and with the help of numerical simulation of combined-cycle plant utilization triple cycle, where the lower cycle is an organic Rankine cycle. Base "REFPROP" describes and allows to calculate the thermodynamic and thermophysical parameters of most of the main substances used in production processes. On the basis of scientific publications on the use of working fluids in an organic Rankine cycle analysis were selected ozone-friendly low-boiling substances: ammonia, butane, pentane and Freon: R134a, R152a, R236fa and R245fa. For these substances have been identified and tabulated molecular weight, temperature of the triple point, boiling point, at atmospheric pressure, the parameters of the critical point, the value of the derivative of the temperature on the entropy of the saturated vapor line and the potential ozone depletion and global warming. It was also identified and tabulated thermodynamic and thermophysical parameters of the steam and liquid substances in a state of saturation at a temperature of 15 °C. This temperature is adopted as the minimum temperature of heat removal in the Rankine cycle when working on the water. Studies have shown that the best thermodynamic, thermal and environmental properties of the considered substances are pentane, butane and R245fa. For a more thorough analysis based on a gas turbine plant NK-36ST it has developed a mathematical model of combined cycle gas turbine (CCGT) triple cycle, where the lower cycle is an organic Rankine cycle, and is used as the air cooler condenser. Air condenser allows stating material at a temperature below 0 °C. Calculation of the

  2. Organization Development: Strategies and Models.

    ERIC Educational Resources Information Center

    Beckhard, Richard

    This book, written for managers, specialists, and students of management, is based largely on the author's experience in helping organization leaders with planned-change efforts, and on related experience of colleagues in the field. Chapter 1 presents the background and causes for the increased concern with organization development and planned…

  3. Modeling and Measuring Organization Capital

    ERIC Educational Resources Information Center

    Atkeson, Andrew; Kehoe, Patrick J.

    2005-01-01

    Manufacturing plants have a clear life cycle: they are born small, grow substantially with age, and eventually die. Economists have long thought that this life cycle is driven by organization capital, the accumulation of plant-specific knowledge. The location of plants in the life cycle determines the size of the payments, or organization rents,…

  4. Guidelines for the Effective Use of Entity-Attribute-Value Modeling for Biomedical Databases

    PubMed Central

    Dinu, Valentin; Nadkarni, Prakash

    2007-01-01

    Purpose To introduce the goals of EAV database modeling, to describe the situations where Entity-Attribute-Value (EAV) modeling is a useful alternative to conventional relational methods of database modeling, and to describe the fine points of implementation in production systems. Methods We analyze the following circumstances: 1) data are sparse and have a large number of applicable attributes, but only a small fraction will apply to a given entity; 2) numerous classes of data need to be represented, each class has a limited number of attributes, but the number of instances of each class is very small. We also consider situations calling for a mixed approach where both conventional and EAV design are used for appropriate data classes. Results and Conclusions In robust production systems, EAV-modeled databases trade a modest data sub-schema for a complex metadata sub-schema. The need to design the metadata effectively makes EAV design potentially more challenging than conventional design. PMID:17098467

  5. FOAM (Functional Ontology Assignments for Metagenomes): A Hidden Markov Model (HMM) database with environmental focus

    DOE PAGESBeta

    Prestat, Emmanuel; David, Maude M.; Hultman, Jenni; Ta , Neslihan; Lamendella, Regina; Dvornik, Jill; Mackelprang, Rachel; Myrold, David D.; Jumpponen, Ari; Tringe, Susannah G.; et al

    2014-09-26

    A new functional gene database, FOAM (Functional Ontology Assignments for Metagenomes), was developed to screen environmental metagenomic sequence datasets. FOAM provides a new functional ontology dedicated to classify gene functions relevant to environmental microorganisms based on Hidden Markov Models (HMMs). Sets of aligned protein sequences (i.e. ‘profiles’) were tailored to a large group of target KEGG Orthologs (KOs) from which HMMs were trained. The alignments were checked and curated to make them specific to the targeted KO. Within this process, sequence profiles were enriched with the most abundant sequences available to maximize the yield of accurate classifier models. An associatedmore » functional ontology was built to describe the functional groups and hierarchy. FOAM allows the user to select the target search space before HMM-based comparison steps and to easily organize the results into different functional categories and subcategories. FOAM is publicly available at http://portal.nersc.gov/project/m1317/FOAM/.« less

  6. Tree-Structured Digital Organisms Model

    NASA Astrophysics Data System (ADS)

    Suzuki, Teruhiko; Nobesawa, Shiho; Tahara, Ikuo

    Tierra and Avida are well-known models of digital organisms. They describe a life process as a sequence of computation codes. A linear sequence model may not be the only way to describe a digital organism, though it is very simple for a computer-based model. Thus we propose a new digital organism model based on a tree structure, which is rather similar to the generic programming. With our model, a life process is a combination of various functions, as if life in the real world is. This implies that our model can easily describe the hierarchical structure of life, and it can simulate evolutionary computation through mutual interaction of functions. We verified our model by simulations that our model can be regarded as a digital organism model according to its definitions. Our model even succeeded in creating species such as viruses and parasites.

  7. Modeling and database for melt-water interfacial heat transfer

    SciTech Connect

    Farmer, M.T.; Spencer, B.W. ); Schneider, J.P. ); Bonomo, B. ); Theofanous, G. )

    1992-01-01

    A mechanistic model is developed to predict the transition superficial gas velocity between bulk cooldown and crust-limited heat transfer regimes in a sparged molten pool with a coolant overlayer. The model has direct applications in the analysis of ex-vessel severe accidents, where molten corium interacts with concrete, thereby producing sparging concrete decomposition gases. The analysis approach embodies thermal, mechanical, and hydrodynamic aspects associated with incipient crust formation at the melt/coolant interface. The model is validated against experiment data obtained with water (melt) and liquid nitrogen (coolant) simulants. Predictions are then made for the critical gas velocity at which crust formation will occur for core material interacting with concrete in the presence of water.

  8. Justifying Database Normalization: A Cost/Benefit Model.

    ERIC Educational Resources Information Center

    Lee, Heeseok

    1995-01-01

    Proposes a cost/benefit model coupled with a decision tree for determining normal forms, which are used in information systems development processes to group data into well-refined structures. The three primary variables that impact the benefits and costs of normalization (reduced anomalies, storage requirements, and transaction response times)…

  9. Circulation Control Model Experimental Database for CFD Validation

    NASA Technical Reports Server (NTRS)

    Paschal, Keith B.; Neuhart, Danny H.; Beeler, George B.; Allan, Brian G.

    2012-01-01

    A 2D circulation control wing was tested in the Basic Aerodynamic Research Tunnel at the NASA Langley Research Center. A traditional circulation control wing employs tangential blowing along the span over a trailing-edge Coanda surface for the purpose of lift augmentation. This model has been tested extensively at the Georgia Tech Research Institute for the purpose of performance documentation at various blowing rates. The current study seeks to expand on the previous work by documenting additional flow-field data needed for validation of computational fluid dynamics. Two jet momentum coefficients were tested during this entry: 0.047 and 0.114. Boundary-layer transition was investigated and turbulent boundary layers were established on both the upper and lower surfaces of the model. Chordwise and spanwise pressure measurements were made, and tunnel sidewall pressure footprints were documented. Laser Doppler Velocimetry measurements were made on both the upper and lower surface of the model at two chordwise locations (x/c = 0.8 and 0.9) to document the state of the boundary layers near the spanwise blowing slot.

  10. Extracting protein alignment models from the sequence database.

    PubMed Central

    Neuwald, A F; Liu, J S; Lipman, D J; Lawrence, C E

    1997-01-01

    Biologists often gain structural and functional insights into a protein sequence by constructing a multiple alignment model of the family. Here a program called Probe fully automates this process of model construction starting from a single sequence. Central to this program is a powerful new method to locate and align only those, often subtly, conserved patterns essential to the family as a whole. When applied to randomly chosen proteins, Probe found on average about four times as many relationships as a pairwise search and yielded many new discoveries. These include: an obscure subfamily of globins in the roundworm Caenorhabditis elegans ; two new superfamilies of metallohydrolases; a lipoyl/biotin swinging arm domain in bacterial membrane fusion proteins; and a DH domain in the yeast Bud3 and Fus2 proteins. By identifying distant relationships and merging families into superfamilies in this way, this analysis further confirms the notion that proteins evolved from relatively few ancient sequences. Moreover, this method automatically generates models of these ancient conserved regions for rapid and sensitive screening of sequences. PMID:9108146

  11. An Object-Relational Ifc Storage Model Based on Oracle Database

    NASA Astrophysics Data System (ADS)

    Li, Hang; Liu, Hua; Liu, Yong; Wang, Yuan

    2016-06-01

    With the building models are getting increasingly complicated, the levels of collaboration across professionals attract more attention in the architecture, engineering and construction (AEC) industry. In order to adapt the change, buildingSMART developed Industry Foundation Classes (IFC) to facilitate the interoperability between software platforms. However, IFC data are currently shared in the form of text file, which is defective. In this paper, considering the object-based inheritance hierarchy of IFC and the storage features of different database management systems (DBMS), we propose a novel object-relational storage model that uses Oracle database to store IFC data. Firstly, establish the mapping rules between data types in IFC specification and Oracle database. Secondly, design the IFC database according to the relationships among IFC entities. Thirdly, parse the IFC file and extract IFC data. And lastly, store IFC data into corresponding tables in IFC database. In experiment, three different building models are selected to demonstrate the effectiveness of our storage model. The comparison of experimental statistics proves that IFC data are lossless during data exchange.

  12. Modeling Powered Aerodynamics for the Orion Launch Abort Vehicle Aerodynamic Database

    NASA Technical Reports Server (NTRS)

    Chan, David T.; Walker, Eric L.; Robinson, Philip E.; Wilson, Thomas M.

    2011-01-01

    Modeling the aerodynamics of the Orion Launch Abort Vehicle (LAV) has presented many technical challenges to the developers of the Orion aerodynamic database. During a launch abort event, the aerodynamic environment around the LAV is very complex as multiple solid rocket plumes interact with each other and the vehicle. It is further complicated by vehicle separation events such as between the LAV and the launch vehicle stack or between the launch abort tower and the crew module. The aerodynamic database for the LAV was developed mainly from wind tunnel tests involving powered jet simulations of the rocket exhaust plumes, supported by computational fluid dynamic simulations. However, limitations in both methods have made it difficult to properly capture the aerodynamics of the LAV in experimental and numerical simulations. These limitations have also influenced decisions regarding the modeling and structure of the aerodynamic database for the LAV and led to compromises and creative solutions. Two database modeling approaches are presented in this paper (incremental aerodynamics and total aerodynamics), with examples showing strengths and weaknesses of each approach. In addition, the unique problems presented to the database developers by the large data space required for modeling a launch abort event illustrate the complexities of working with multi-dimensional data.

  13. The Cambridge MRI database for animal models of Huntington disease.

    PubMed

    Sawiak, Stephen J; Morton, A Jennifer

    2016-01-01

    We describe the Cambridge animal brain magnetic resonance imaging repository comprising 400 datasets to date from mouse models of Huntington disease. The data include raw images as well as segmented grey and white matter images with maps of cortical thickness. All images and phenotypic data for each subject are freely-available without restriction from (http://www.dspace.cam.ac.uk/handle/1810/243361/). Software and anatomical population templates optimised for animal brain analysis with MRI are also available from this site.

  14. Database Design Learning: A Project-Based Approach Organized through a Course Management System

    ERIC Educational Resources Information Center

    Dominguez, Cesar; Jaime, Arturo

    2010-01-01

    This paper describes an active method for database design learning through practical tasks development by student teams in a face-to-face course. This method integrates project-based learning, and project management techniques and tools. Some scaffolding is provided at the beginning that forms a skeleton that adapts to a great variety of…

  15. Hydraulic fracture propagation modeling and data-based fracture identification

    NASA Astrophysics Data System (ADS)

    Zhou, Jing

    Successful shale gas and tight oil production is enabled by the engineering innovation of horizontal drilling and hydraulic fracturing. Hydraulically induced fractures will most likely deviate from the bi-wing planar pattern and generate complex fracture networks due to mechanical interactions and reservoir heterogeneity, both of which render the conventional fracture simulators insufficient to characterize the fractured reservoir. Moreover, in reservoirs with ultra-low permeability, the natural fractures are widely distributed, which will result in hydraulic fractures branching and merging at the interface and consequently lead to the creation of more complex fracture networks. Thus, developing a reliable hydraulic fracturing simulator, including both mechanical interaction and fluid flow, is critical in maximizing hydrocarbon recovery and optimizing fracture/well design and completion strategy in multistage horizontal wells. A novel fully coupled reservoir flow and geomechanics model based on the dual-lattice system is developed to simulate multiple nonplanar fractures' propagation in both homogeneous and heterogeneous reservoirs with or without pre-existing natural fractures. Initiation, growth, and coalescence of the microcracks will lead to the generation of macroscopic fractures, which is explicitly mimicked by failure and removal of bonds between particles from the discrete element network. This physics-based modeling approach leads to realistic fracture patterns without using the empirical rock failure and fracture propagation criteria required in conventional continuum methods. Based on this model, a sensitivity study is performed to investigate the effects of perforation spacing, in-situ stress anisotropy, rock properties (Young's modulus, Poisson's ratio, and compressive strength), fluid properties, and natural fracture properties on hydraulic fracture propagation. In addition, since reservoirs are buried thousands of feet below the surface, the

  16. Transport and Environment Database System (TRENDS): Maritime air pollutant emission modelling

    NASA Astrophysics Data System (ADS)

    Georgakaki, Aliki; Coffey, Robert A.; Lock, Graham; Sorenson, Spencer C.

    This paper reports the development of the maritime module within the framework of the Transport and Environment Database System (TRENDS) project. A detailed database has been constructed for the calculation of energy consumption and air pollutant emissions. Based on an in-house database of commercial vessels kept at the Technical University of Denmark, relationships between the fuel consumption and size of different vessels have been developed, taking into account the fleet's age and service speed. The technical assumptions and factors incorporated in the database are presented, including changes from findings reported in Methodologies for Estimating air pollutant Emissions from Transport (MEET). The database operates on statistical data provided by Eurostat, which describe vessel and freight movements from and towards EU 15 major ports. Data are at port to Maritime Coastal Area (MCA) level, so a bottom-up approach is used. A port to MCA distance database has also been constructed for the purpose of the study. This was the first attempt to use Eurostat maritime statistics for emission modelling; and the problems encountered, since the statistical data collection was not undertaken with a view to this purpose, are mentioned. Examples of the results obtained by the database are presented. These include detailed air pollutant emission calculations for bulk carriers entering the port of Helsinki, as an example of the database operation, and aggregate results for different types of movements for France. Overall estimates of SO x and NO x emission caused by shipping traffic between the EU 15 countries are in the area of 1 and 1.5 million tonnes, respectively.

  17. Query Monitoring and Analysis for Database Privacy - A Security Automata Model Approach

    PubMed Central

    Kumar, Anand; Ligatti, Jay; Tu, Yi-Cheng

    2015-01-01

    Privacy and usage restriction issues are important when valuable data are exchanged or acquired by different organizations. Standard access control mechanisms either restrict or completely grant access to valuable data. On the other hand, data obfuscation limits the overall usability and may result in loss of total value. There are no standard policy enforcement mechanisms for data acquired through mutual and copyright agreements. In practice, many different types of policies can be enforced in protecting data privacy. Hence there is the need for an unified framework that encapsulates multiple suites of policies to protect the data. We present our vision of an architecture named security automata model (SAM) to enforce privacy-preserving policies and usage restrictions. SAM analyzes the input queries and their outputs to enforce various policies, liberating data owners from the burden of monitoring data access. SAM allows administrators to specify various policies and enforces them to monitor queries and control the data access. Our goal is to address the problems of data usage control and protection through privacy policies that can be defined, enforced, and integrated with the existing access control mechanisms using SAM. In this paper, we lay out the theoretical foundation of SAM, which is based on an automata named Mandatory Result Automata. We also discuss the major challenges of implementing SAM in a real-world database environment as well as ideas to meet such challenges. PMID:26997936

  18. Data model and relational database design for the New Jersey Water-Transfer Data System (NJWaTr)

    USGS Publications Warehouse

    Tessler, Steven

    2003-01-01

    The New Jersey Water-Transfer Data System (NJWaTr) is a database design for the storage and retrieval of water-use data. NJWaTr can manage data encompassing many facets of water use, including (1) the tracking of various types of water-use activities (withdrawals, returns, transfers, distributions, consumptive-use, wastewater collection, and treatment); (2) the storage of descriptions, classifications and locations of places and organizations involved in water-use activities; (3) the storage of details about measured or estimated volumes of water associated with water-use activities; and (4) the storage of information about data sources and water resources associated with water use. In NJWaTr, each water transfer occurs unidirectionally between two site objects, and the sites and conveyances form a water network. The core entities in the NJWaTr model are site, conveyance, transfer/volume, location, and owner. Other important entities include water resource (used for withdrawals and returns), data source, permit, and alias. Multiple water-exchange estimates based on different methods or data sources can be stored for individual transfers. Storage of user-defined details is accommodated for several of the main entities. Many tables contain classification terms to facilitate the detailed description of data items and can be used for routine or custom data summarization. NJWaTr accommodates single-user and aggregate-user water-use data, can be used for large or small water-network projects, and is available as a stand-alone Microsoft? Access database. Data stored in the NJWaTr structure can be retrieved in user-defined combinations to serve visualization and analytical applications. Users can customize and extend the database, link it to other databases, or implement the design in other relational database applications.

  19. TogoTable: cross-database annotation system using the Resource Description Framework (RDF) data model

    PubMed Central

    Kawano, Shin; Watanabe, Tsutomu; Mizuguchi, Sohei; Araki, Norie; Katayama, Toshiaki; Yamaguchi, Atsuko

    2014-01-01

    TogoTable (http://togotable.dbcls.jp/) is a web tool that adds user-specified annotations to a table that a user uploads. Annotations are drawn from several biological databases that use the Resource Description Framework (RDF) data model. TogoTable uses database identifiers (IDs) in the table as a query key for searching. RDF data, which form a network called Linked Open Data (LOD), can be searched from SPARQL endpoints using a SPARQL query language. Because TogoTable uses RDF, it can integrate annotations from not only the reference database to which the IDs originally belong, but also externally linked databases via the LOD network. For example, annotations in the Protein Data Bank can be retrieved using GeneID through links provided by the UniProt RDF. Because RDF has been standardized by the World Wide Web Consortium, any database with annotations based on the RDF data model can be easily incorporated into this tool. We believe that TogoTable is a valuable Web tool, particularly for experimental biologists who need to process huge amounts of data such as high-throughput experimental output. PMID:24829452

  20. Microporoelastic Modeling of Organic-Rich Shales

    NASA Astrophysics Data System (ADS)

    Khosh Sokhan Monfared, S.; Abedi, S.; Ulm, F. J.

    2014-12-01

    Organic-rich shale is an extremely complex, naturally occurring geo-composite. The heterogeneous nature of organic-rich shale and its anisotropic behavior pose grand challenges for characterization, modeling and engineering design The intricacy of organic-rich shale, in the context of its mechanical and poromechanical properties, originates in the presence of organic/inorganic constituents and their interfaces as well as the occurrence of porosity and elastic anisotropy, at multiple length scales. To capture the contributing mechanisms, of 1st order, responsible for organic-rich shale complex behavior, we introduce an original approach for micromechanical modeling of organic-rich shales which accounts for the effect of maturity of organics on the overall elasticity through morphology considerations. This morphology contribution is captured by means of an effective media theory that bridges the gap between immature and mature systems through the choice of system's microtexture; namely a matrix-inclusion morphology (Mori-Tanaka) for immature systems and a polycrystal/granular morphology for mature systems. Also, we show that interfaces play a role on the effective elasticity of mature, organic-rich shales. The models are calibrated by means of ultrasonic pulse velocity measurements of elastic properties and validated by means of nanoindentation results. Sensitivity analyses using Spearman's Partial Rank Correlation Coefficient shows the importance of porosity and Total Organic Carbon (TOC) as key input parameters for accurate model predictions. These modeling developments pave the way to reach a "unique" set of clay properties and highlight the importance of depositional environment, burial and diagenetic processes on overall mechanical and poromechanical behavior of organic-rich shale. These developments also emphasize the importance of understanding and modeling clay elasticity and organic maturity on the overall rock behavior which is of critical importance for a

  1. Other model organisms for sarcomeric muscle diseases

    PubMed Central

    Sparrow, John; Hughes, Simon M.; Segalat, Laurent

    2012-01-01

    Model organisms are vital to our understanding of human muscle biology and disease. The potential of the nematode Caenorhabditis elegans, the fruitfly, Drosophila melanogaster and the zebrafish, Danio rerio, as model genetic organisms for the study of human muscle disease is discussed by examining their muscle biology, muscle genetics and development. The powerful genetic tools available with each organism are outlined. It is concluded that these organisms have already demonstrated potential in facilitating the study of muscle disease and in screening for therapeutic agents. PMID:19181102

  2. Modeling the High Speed Research Cycle 2B Longitudinal Aerodynamic Database Using Multivariate Orthogonal Functions

    NASA Technical Reports Server (NTRS)

    Morelli, E. A.; Proffitt, M. S.

    1999-01-01

    The data for longitudinal non-dimensional, aerodynamic coefficients in the High Speed Research Cycle 2B aerodynamic database were modeled using polynomial expressions identified with an orthogonal function modeling technique. The discrepancy between the tabular aerodynamic data and the polynomial models was tested and shown to be less than 15 percent for drag, lift, and pitching moment coefficients over the entire flight envelope. Most of this discrepancy was traced to smoothing local measurement noise and to the omission of mass case 5 data in the modeling process. A simulation check case showed that the polynomial models provided a compact and accurate representation of the nonlinear aerodynamic dependencies contained in the HSR Cycle 2B tabular aerodynamic database.

  3. Integrated Functional and Executional Modelling of Software Using Web-Based Databases

    NASA Technical Reports Server (NTRS)

    Kulkarni, Deepak; Marietta, Roberta

    1998-01-01

    NASA's software subsystems undergo extensive modification and updates over the operational lifetimes. It is imperative that modified software should satisfy safety goals. This report discusses the difficulties encountered in doing so and discusses a solution based on integrated modelling of software, use of automatic information extraction tools, web technology and databases.

  4. A genome-scale metabolic flux model of Escherichia coli K–12 derived from the EcoCyc database

    PubMed Central

    2014-01-01

    advantages can be derived from the combination of model organism databases and flux balance modeling represented by MetaFlux. Interpretation of the EcoCyc database as a flux balance model results in a highly accurate metabolic model and provides a rigorous consistency check for information stored in the database. PMID:24974895

  5. Genome databases

    SciTech Connect

    Courteau, J.

    1991-10-11

    Since the Genome Project began several years ago, a plethora of databases have been developed or are in the works. They range from the massive Genome Data Base at Johns Hopkins University, the central repository of all gene mapping information, to small databases focusing on single chromosomes or organisms. Some are publicly available, others are essentially private electronic lab notebooks. Still others limit access to a consortium of researchers working on, say, a single human chromosome. An increasing number incorporate sophisticated search and analytical software, while others operate as little more than data lists. In consultation with numerous experts in the field, a list has been compiled of some key genome-related databases. The list was not limited to map and sequence databases but also included the tools investigators use to interpret and elucidate genetic data, such as protein sequence and protein structure databases. Because a major goal of the Genome Project is to map and sequence the genomes of several experimental animals, including E. coli, yeast, fruit fly, nematode, and mouse, the available databases for those organisms are listed as well. The author also includes several databases that are still under development - including some ambitious efforts that go beyond data compilation to create what are being called electronic research communities, enabling many users, rather than just one or a few curators, to add or edit the data and tag it as raw or confirmed.

  6. The LAILAPS search engine: a feature model for relevance ranking in life science databases.

    PubMed

    Lange, Matthias; Spies, Karl; Colmsee, Christian; Flemming, Steffen; Klapperstück, Matthias; Scholz, Uwe

    2010-03-25

    Efficient and effective information retrieval in life sciences is one of the most pressing challenge in bioinformatics. The incredible growth of life science databases to a vast network of interconnected information systems is to the same extent a big challenge and a great chance for life science research. The knowledge found in the Web, in particular in life-science databases, are a valuable major resource. In order to bring it to the scientist desktop, it is essential to have well performing search engines. Thereby, not the response time nor the number of results is important. The most crucial factor for millions of query results is the relevance ranking. In this paper, we present a feature model for relevance ranking in life science databases and its implementation in the LAILAPS search engine. Motivated by the observation of user behavior during their inspection of search engine result, we condensed a set of 9 relevance discriminating features. These features are intuitively used by scientists, who briefly screen database entries for potential relevance. The features are both sufficient to estimate the potential relevance, and efficiently quantifiable. The derivation of a relevance prediction function that computes the relevance from this features constitutes a regression problem. To solve this problem, we used artificial neural networks that have been trained with a reference set of relevant database entries for 19 protein queries. Supporting a flexible text index and a simple data import format, this concepts are implemented in the LAILAPS search engine. It can easily be used both as search engine for comprehensive integrated life science databases and for small in-house project databases. LAILAPS is publicly available for SWISSPROT data at http://lailaps.ipk-gatersleben.de.

  7. A unified timeline model and user interface for multimedia medical databases.

    PubMed

    Dionisio, J D; Cárdenas, A F; Taira, R K; Aberle, D R; Chu, W W; McNitt-Gray, M F; Goldin, J; Lufkin, R B

    1996-01-01

    A multimedia medical database model and prototype is described for supporting a timeline-based presentation of information. The database links image and text data in a way that permits users to look at medical information in a single unified view. Various visualization programs permit the user to view data in various ways, including full image views, graphs, and tables. Our technology is applied for proof-of-concept to two areas: thoracic oncology and thermal tumor ablation therapy of the brain. This effort is part of the multidisciplinary KMeD project in collaboration with medical research and clinical treatment projects at UCLA.

  8. Gas Chromatography and Mass Spectrometry Measurements and Protocols for Database and Library Development Relating to Organic Species in Support of the Mars Science Laboratory

    NASA Astrophysics Data System (ADS)

    Misra, P.; Garcia, R.; Mahaffy, P. R.

    2010-04-01

    An organic contaminant database and library has been developed for use with the Sample Analysis at Mars (SAM) instrumentation utilizing laboratory-based Gas Chromatography-Mass Spectrometry measurements of pyrolyzed and baked material samples.

  9. ITVO and BaSTI: databases and services for cosmological and stellar models.

    NASA Astrophysics Data System (ADS)

    Manzato, P.; Molinaro, M.; Gasparo, F.; Pasian, F.; Pietrinferni, A.; Cassisi, S.; Gheller, C.; Ameglio, S.; Murante, G.; Borgani, S.

    We have created a database structure to store the metadata of different types of cosmological simulations (Gadget, Enzo, FLY) and the first relational database for stellar evolution models BaSTI, it includes tracks and isochrones computed with the FRANEC code. We are also studying the feasibility of including different sets of theory data and services in the Virtual Observatory (VObs). Some examples of services are: the calculation on-the-fly of the profiles of some quantities for the simulated galaxy clusters, the preview of the object image opened with a VObs tool and retrieve a VOTable standard format. Furthermore, the BaSTI database development is the use case for studying the feasibility of storing in it the output of new simulations performed using the Grid infrastructure as demonstrating in the VO-DCA WP5, EU funded project. All that could be matter of discussion between the tool developers and the users, the scientists.

  10. Database on Pathogen and Indicator Organism Survival in Soils and other Environmental Media

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Data on survival of pathogen and indicator organism in soils, sediments, organic waste, and waters represent the key information for evaluating management practices and predicting fate and transport of the microorganisms. Such data are, in general, available, but are spread across thousands of publi...

  11. Creating a model to detect dairy cattle farms with poor welfare using a national database.

    PubMed

    Krug, C; Haskell, M J; Nunes, T; Stilwell, G

    2015-12-01

    The objective of this study was to determine whether dairy farms with poor cow welfare could be identified using a national database for bovine identification and registration that monitors cattle deaths and movements. The welfare of dairy cattle was assessed using the Welfare Quality(®) protocol (WQ) on 24 Portuguese dairy farms and on 1930 animals. Five farms were classified as having poor welfare and the other 19 were classified as having good welfare. Fourteen million records from the national cattle database were analysed to identify potential welfare indicators for dairy farms. Fifteen potential national welfare indicators were calculated based on that database, and the link between the results on the WQ evaluation and the national cattle database was made using the identification code of each farm. Within the potential national welfare indicators, only two were significantly different between farms with good welfare and poor welfare, 'proportion of on-farm deaths' (p<0.01) and 'female/male birth ratio' (p<0.05). To determine whether the database welfare indicators could be used to distinguish farms with good welfare from farms with poor welfare, we created a model using the classifier J48 of Waikato Environment for Knowledge Analysis. The model was a decision tree based on two variables, 'proportion of on-farm deaths' and 'calving-to-calving interval', and it was able to correctly identify 70% and 79% of the farms classified as having poor and good welfare, respectively. The national cattle database analysis could be useful in helping official veterinary services in detecting farms that have poor welfare and also in determining which welfare indicators are poor on each particular farm.

  12. Mouse genome database 2016.

    PubMed

    Bult, Carol J; Eppig, Janan T; Blake, Judith A; Kadin, James A; Richardson, Joel E

    2016-01-01

    The Mouse Genome Database (MGD; http://www.informatics.jax.org) is the primary community model organism database for the laboratory mouse and serves as the source for key biological reference data related to mouse genes, gene functions, phenotypes and disease models with a strong emphasis on the relationship of these data to human biology and disease. As the cost of genome-scale sequencing continues to decrease and new technologies for genome editing become widely adopted, the laboratory mouse is more important than ever as a model system for understanding the biological significance of human genetic variation and for advancing the basic research needed to support the emergence of genome-guided precision medicine. Recent enhancements to MGD include new graphical summaries of biological annotations for mouse genes, support for mobile access to the database, tools to support the annotation and analysis of sets of genes, and expanded support for comparative biology through the expansion of homology data.

  13. Mouse genome database 2016

    PubMed Central

    Bult, Carol J.; Eppig, Janan T.; Blake, Judith A.; Kadin, James A.; Richardson, Joel E.

    2016-01-01

    The Mouse Genome Database (MGD; http://www.informatics.jax.org) is the primary community model organism database for the laboratory mouse and serves as the source for key biological reference data related to mouse genes, gene functions, phenotypes and disease models with a strong emphasis on the relationship of these data to human biology and disease. As the cost of genome-scale sequencing continues to decrease and new technologies for genome editing become widely adopted, the laboratory mouse is more important than ever as a model system for understanding the biological significance of human genetic variation and for advancing the basic research needed to support the emergence of genome-guided precision medicine. Recent enhancements to MGD include new graphical summaries of biological annotations for mouse genes, support for mobile access to the database, tools to support the annotation and analysis of sets of genes, and expanded support for comparative biology through the expansion of homology data. PMID:26578600

  14. Mouse genome database 2016.

    PubMed

    Bult, Carol J; Eppig, Janan T; Blake, Judith A; Kadin, James A; Richardson, Joel E

    2016-01-01

    The Mouse Genome Database (MGD; http://www.informatics.jax.org) is the primary community model organism database for the laboratory mouse and serves as the source for key biological reference data related to mouse genes, gene functions, phenotypes and disease models with a strong emphasis on the relationship of these data to human biology and disease. As the cost of genome-scale sequencing continues to decrease and new technologies for genome editing become widely adopted, the laboratory mouse is more important than ever as a model system for understanding the biological significance of human genetic variation and for advancing the basic research needed to support the emergence of genome-guided precision medicine. Recent enhancements to MGD include new graphical summaries of biological annotations for mouse genes, support for mobile access to the database, tools to support the annotation and analysis of sets of genes, and expanded support for comparative biology through the expansion of homology data. PMID:26578600

  15. High rate of unemployment after liver transplantation: analysis of the United Network for Organ Sharing database.

    PubMed

    Huda, Amina; Newcomer, Robert; Harrington, Charlene; Blegen, Mary G; Keeffe, Emmet B

    2012-01-01

    The goal of liver transplantation (LT) is to maximize the length and quality of a patient's life and facilitate his or her return to full productivity. The aims of this study were (1) to use the United Network for Organ Sharing (UNOS) data set to determine the proportions of recipients who were employed and unemployed within 24 months after LT between 2002 and 2008 and (2) to examine the factors associated with a return to employment. UNOS data that were collected since the adoption of the Model for End-Stage Liver Disease scoring system on February 27, 2002 were analyzed. There were 21,942 transplant recipients who met the inclusion criteria. The employment status of the recipients was analyzed within a 60-day window at the following times after transplantation: 6, 12, and 24 months. Approximately one-quarter of the LT recipients (5360 or 24.4%) were employed within 24 months after transplantation, and the remaining recipients had not returned to work. The demographic variables that were independently associated with posttransplant employment included an age of 18 to 40 years, male sex, a college degree, Caucasian race, and pretransplant employment. Patients with alcoholic liver disease had a significantly lower rate of employment than patients with other etiologies of liver disease. The recipients who were employed after transplantation had significantly better functional status than those who were not employed. In conclusion, the employment rate after LT is low, with only one-quarter of LT recipients employed. New national and individual transplant program policies are needed to assess the root causes of unemployment in recipients who wish to work after LT.

  16. D Digital Model Database Applied to Conservation and Research of Wooden Construction in China

    NASA Astrophysics Data System (ADS)

    Zheng, Y.

    2013-07-01

    Protected by the Tai-Hang Mountains, Shanxi Province, located in north central China, is a highly prosperous, densely populated valley and considered to be one of the cradles of Chinese civilization. Its continuous habitation and rich culture have given rise to a large number of temple complexes and pavilions. Among these structures, 153 can be dated as early as from the Tang dynasty (618- 907C.E.) to the end of the Yuan dynasty (1279-1368C.E.) in Southern Shanxi area. The buildings are the best-preserved examples of wooden Chinese architecture in existence, exemplifying historic building technology and displaying highly intricate architectural decoration and detailing. They have survived war, earthquakes, and, in the last hundred years, neglect. In 2005, a decade-long conservation project was initiated by the State Administration of Cultural Heritage of China (SACH) to conserve and document these important buildings. The conservation process requires stabilization, conservation of important features, and, where necessary, partial dismantlement in order to replace unsound structural elements. Project team of CHCC have developed a practical recording system that created a record of all building components prior to and during the conservation process. After that we are trying to establish a comprehensive database which include all of the 153 earlier buildings, through which we can easily entering, browse, indexing information of the wooden construction, even deep into component details. The Database can help us to carry out comparative studies of these wooden structures, and, provide important support for the continued conservation of these heritage buildings. For some of the most important wooden structure, we have established three-dimensional models. Connected the Database with 3D Digital Model based on ArcGIS, we have developed 3D Digital Model Database for these cherish buildings. The 3D Digital Model Database helps us set up an integrate information inventory

  17. Photochemistry of Model Organic Aerosol Systems

    NASA Astrophysics Data System (ADS)

    Mang, S. A.; Bateman, A. P.; Dailo, M.; Do, T.; Nizkorodov, S. A.; Pan, X.; Underwood, J. S.; Walser, M. L.

    2007-05-01

    Up to 90 percent of urban aerosol particles have been shown to contain organic molecules. Reactions of these particles with atmospheric oxidants and/or sunlight result in large changes in their composition, toxicity, and ability to act as cloud condensation nuclei. For this reason, chemistry of model organic aerosol particles initiated by oxidation and direct photolysis is of great interest to atmospheric, climate, and health scientists. Most studies in this area have focused on identifying the products of oxidation of the organic aerosols, while the products of direct photolysis of the resulting molecules remaining in the aerosol particle have been left mostly unexplored. We have explored direct photolytic processes occurring in selected organic aerosol systems using infrared cavity ringdown spectroscopy to identify small gas phase products of photolysis, and mass-spectrometric and photometric techniques to study the condensed phase products. The first model system was secondary organic aerosol formed from the oxidation of several monoterpenes by ozone in the presence and absence of NOx, under different humidities. The second system modeled after oxidatively aged primary organic aerosol particles was a thin film of either alkanes or saturated fatty acids oxidized in several different ways, with the oxidation initiated by ozone, chlorine atom, or OH. In every case, the general conclusion was that the photochemical processing of model organic aerosols is significant. Such direct photolysis processes are believed to age organic aerosol particles on time scales that are short compared to the particles' atmospheric lifetimes.

  18. Significant speedup of database searches with HMMs by search space reduction with PSSM family models

    PubMed Central

    Beckstette, Michael; Homann, Robert; Giegerich, Robert; Kurtz, Stefan

    2009-01-01

    Motivation: Profile hidden Markov models (pHMMs) are currently the most popular modeling concept for protein families. They provide sensitive family descriptors, and sequence database searching with pHMMs has become a standard task in today's genome annotation pipelines. On the downside, searching with pHMMs is computationally expensive. Results: We propose a new method for efficient protein family classification and for speeding up database searches with pHMMs as is necessary for large-scale analysis scenarios. We employ simpler models of protein families called position-specific scoring matrices family models (PSSM-FMs). For fast database search, we combine full-text indexing, efficient exact p-value computation of PSSM match scores and fast fragment chaining. The resulting method is well suited to prefilter the set of sequences to be searched for subsequent database searches with pHMMs. We achieved a classification performance only marginally inferior to hmmsearch, yet, results could be obtained in a fraction of runtime with a speedup of >64-fold. In experiments addressing the method's ability to prefilter the sequence space for subsequent database searches with pHMMs, our method reduces the number of sequences to be searched with hmmsearch to only 0.80% of all sequences. The filter is very fast and leads to a total speedup of factor 43 over the unfiltered search, while retaining >99.5% of the original results. In a lossless filter setup for hmmsearch on UniProtKB/Swiss-Prot, we observed a speedup of factor 92. Availability: The presented algorithms are implemented in the program PoSSuMsearch2, available for download at http://bibiserv.techfak.uni-bielefeld.de/possumsearch2/. Contact: beckstette@zbh.uni-hamburg.de Supplementary information: Supplementary data are available at Bioinformatics online. PMID:19828575

  19. Modeling personnel turnover in the parametric organization

    NASA Technical Reports Server (NTRS)

    Dean, Edwin B.

    1991-01-01

    A model is developed for simulating the dynamics of a newly formed organization, credible during all phases of organizational development. The model development process is broken down into the activities of determining the tasks required for parametric cost analysis (PCA), determining the skills required for each PCA task, determining the skills available in the applicant marketplace, determining the structure of the model, implementing the model, and testing it. The model, parameterized by the likelihood of job function transition, has demonstrated by the capability to represent the transition of personnel across functional boundaries within a parametric organization using a linear dynamical system, and the ability to predict required staffing profiles to meet functional needs at the desired time. The model can be extended by revisions of the state and transition structure to provide refinements in functional definition for the parametric and extended organization.

  20. Cardiac Electromechanical Models: From Cell to Organ

    PubMed Central

    Trayanova, Natalia A.; Rice, John Jeremy

    2011-01-01

    The heart is a multiphysics and multiscale system that has driven the development of the most sophisticated mathematical models at the frontiers of computational physiology and medicine. This review focuses on electromechanical (EM) models of the heart from the molecular level of myofilaments to anatomical models of the organ. Because of the coupling in terms of function and emergent behaviors at each level of biological hierarchy, separation of behaviors at a given scale is difficult. Here, a separation is drawn at the cell level so that the first half addresses subcellular/single-cell models and the second half addresses organ models. At the subcellular level, myofilament models represent actin–myosin interaction and Ca-based activation. The discussion of specific models emphasizes the roles of cooperative mechanisms and sarcomere length dependence of contraction force, considered to be the cellular basis of the Frank–Starling law. A model of electrophysiology and Ca handling can be coupled to a myofilament model to produce an EM cell model, and representative examples are summarized to provide an overview of the progression of the field. The second half of the review covers organ-level models that require solution of the electrical component as a reaction–diffusion system and the mechanical component, in which active tension generated by the myocytes produces deformation of the organ as described by the equations of continuum mechanics. As outlined in the review, different organ-level models have chosen to use different ionic and myofilament models depending on the specific application; this choice has been largely dictated by compromises between model complexity and computational tractability. The review also addresses application areas of EM models such as cardiac resynchronization therapy and the role of mechano-electric coupling in arrhythmias and defibrillation. PMID:21886622

  1. [Progress of chlamydomonas as a model organism].

    PubMed

    Xie, Chuan-Xiao; Han, Wei; Yu, Zeng-Liang

    2003-05-01

    The unicellular alga Chlamydomonas offers a simple life cycle, easy culture and isolation of series of mutants, established the techniques and tool kit for molecular genetics and genetics analysis. It is now becoming the model organism for studies on photosynthesis in plant, flagellar assembly and function, cell cycle and circadian rhythms, signal transduction, light perception and cell recognition. It is summarized the progress of study on Chlamydomonas as a model organism in this paper.

  2. Model Organisms for Studying the Cell Cycle.

    PubMed

    Tang, Zhaohua

    2016-01-01

    Regulation of the cell-division cycle is fundamental for the growth, development, and reproduction of all species of life. In the past several decades, a conserved theme of cell cycle regulation has emerged from research in diverse model organisms. A comparison of distinct features of several diverse model organisms commonly used in cell cycle studies highlights their suitability for various experimental approaches, and recaptures their contributions to our current understanding of the eukaryotic cell cycle. A historic perspective presents a recollection of the breakthrough upon unfolding the universal principles of cell cycle control by scientists working with diverse model organisms, thereby appreciating the discovery pathways in this field. A comprehensive understanding is necessary to address current challenging questions about cell cycle control. Advances in genomics, proteomics, quantitative methodologies, and approaches of systems biology are redefining the traditional concept of what constitutes a model organism and have established a new era for development of novel, and refinement of the established model organisms. Researchers working in the field are no longer separated by their favorite model organisms; they have become more integrated into a larger community for gaining greater insights into how a cell divides and cycles. The new technologies provide a broad evolutionary spectrum of the cell-division cycle and allow informative comparisons among different species at a level that has never been possible, exerting unimaginable impact on our comprehensive understanding of cell cycle regulation.

  3. Acquisition of Seco Creek GIS database and it's use in water quality models

    SciTech Connect

    Steers, C.A.; Steiner, M.; Taylor, B. )

    1993-02-01

    The Seco Creek Water Quality Demonstration Project covers 1,700,670 acres in parts of Bandera, Frio, Medina and Uvalde Counties in south central Texas. The Seco Creek Database was constructed as part of the Soil Conservation Service's National Water Quality Program to develop hydrologic tools that measure the effects of agricultural nonpoint source pollution and to demonstrate the usefulness of GIS in natural resources management. This project will be part of a GRASS-Water Quality Model Interface which will incorporate watershed models with water quality planning and implementation by January of 1994. The Seco Creek Demonstration Area is the sole water supply for 1.3 million in the San Antonio Area. The database constructed for the project will help maintain the excellent water quality that flows directly into the Edwards Aquifer. The database consists of several vector and raster layers including: SSURGO quality soils, elevation, roads, streams and detailed data on field ownership, cropping and grazing practices and other landuses. This paper will consist of the development and planned uses of the Seco Creek Database.

  4. Measuring the effects of distributed database models on transaction availability measures

    NASA Technical Reports Server (NTRS)

    Mukkamala, Ravi

    1991-01-01

    Data distribution, data replication, and system reliability are key factors in determining the availability measures for transactions in distributed database systems. In order to simplify the evaluation of these measures, database designers and researchers tend to make unrealistic assumptions about these factors. Here, the effect of such assumptions on the computational complexity and accuracy of such evaluations is investigated. A database system is represented with five parameters related to the above factors. Probabilistic analysis is employed to evaluate the availability of read-one and read-write transactions. Both the read-one/write-all and the majority-read/majority-write replication control policies are considered. It is concluded that transaction availability is more sensitive to variations in degrees of replication, less sensitive to data distribution, and insensitive to reliability variations in a heterogeneous system. The computational complexity of the evaluations is found to be mainly determined by the chosen distributed database model, while the accuracy of the results are not so much dependent on the models.

  5. The Chinchilla Research Resource Database: resource for an otolaryngology disease model.

    PubMed

    Shimoyama, Mary; Smith, Jennifer R; De Pons, Jeff; Tutaj, Marek; Khampang, Pawjai; Hong, Wenzhou; Erbe, Christy B; Ehrlich, Garth D; Bakaletz, Lauren O; Kerschner, Joseph E

    2016-01-01

    The long-tailed chinchilla (Chinchilla lanigera) is an established animal model for diseases of the inner and middle ear, among others. In particular, chinchilla is commonly used to study diseases involving viral and bacterial pathogens and polymicrobial infections of the upper respiratory tract and the ear, such as otitis media. The value of the chinchilla as a model for human diseases prompted the sequencing of its genome in 2012 and the more recent development of the Chinchilla Research Resource Database (http://crrd.mcw.edu) to provide investigators with easy access to relevant datasets and software tools to enhance their research. The Chinchilla Research Resource Database contains a complete catalog of genes for chinchilla and, for comparative purposes, human. Chinchilla genes can be viewed in the context of their genomic scaffold positions using the JBrowse genome browser. In contrast to the corresponding records at NCBI, individual gene reports at CRRD include functional annotations for Disease, Gene Ontology (GO) Biological Process, GO Molecular Function, GO Cellular Component and Pathway assigned to chinchilla genes based on annotations from the corresponding human orthologs. Data can be retrieved via keyword and gene-specific searches. Lists of genes with similar functional attributes can be assembled by leveraging the hierarchical structure of the Disease, GO and Pathway vocabularies through the Ontology Search and Browser tool. Such lists can then be further analyzed for commonalities using the Gene Annotator (GA) Tool. All data in the Chinchilla Research Resource Database is freely accessible and downloadable via the CRRD FTP site or using the download functions available in the search and analysis tools. The Chinchilla Research Resource Database is a rich resource for researchers using, or considering the use of, chinchilla as a model for human disease.Database URL: http://crrd.mcw.edu.

  6. The Chinchilla Research Resource Database: resource for an otolaryngology disease model.

    PubMed

    Shimoyama, Mary; Smith, Jennifer R; De Pons, Jeff; Tutaj, Marek; Khampang, Pawjai; Hong, Wenzhou; Erbe, Christy B; Ehrlich, Garth D; Bakaletz, Lauren O; Kerschner, Joseph E

    2016-01-01

    The long-tailed chinchilla (Chinchilla lanigera) is an established animal model for diseases of the inner and middle ear, among others. In particular, chinchilla is commonly used to study diseases involving viral and bacterial pathogens and polymicrobial infections of the upper respiratory tract and the ear, such as otitis media. The value of the chinchilla as a model for human diseases prompted the sequencing of its genome in 2012 and the more recent development of the Chinchilla Research Resource Database (http://crrd.mcw.edu) to provide investigators with easy access to relevant datasets and software tools to enhance their research. The Chinchilla Research Resource Database contains a complete catalog of genes for chinchilla and, for comparative purposes, human. Chinchilla genes can be viewed in the context of their genomic scaffold positions using the JBrowse genome browser. In contrast to the corresponding records at NCBI, individual gene reports at CRRD include functional annotations for Disease, Gene Ontology (GO) Biological Process, GO Molecular Function, GO Cellular Component and Pathway assigned to chinchilla genes based on annotations from the corresponding human orthologs. Data can be retrieved via keyword and gene-specific searches. Lists of genes with similar functional attributes can be assembled by leveraging the hierarchical structure of the Disease, GO and Pathway vocabularies through the Ontology Search and Browser tool. Such lists can then be further analyzed for commonalities using the Gene Annotator (GA) Tool. All data in the Chinchilla Research Resource Database is freely accessible and downloadable via the CRRD FTP site or using the download functions available in the search and analysis tools. The Chinchilla Research Resource Database is a rich resource for researchers using, or considering the use of, chinchilla as a model for human disease.Database URL: http://crrd.mcw.edu. PMID:27173523

  7. The Chinchilla Research Resource Database: resource for an otolaryngology disease model

    PubMed Central

    Shimoyama, Mary; Smith, Jennifer R.; De Pons, Jeff; Tutaj, Marek; Khampang, Pawjai; Hong, Wenzhou; Erbe, Christy B.; Ehrlich, Garth D.; Bakaletz, Lauren O.; Kerschner, Joseph E.

    2016-01-01

    The long-tailed chinchilla (Chinchilla lanigera) is an established animal model for diseases of the inner and middle ear, among others. In particular, chinchilla is commonly used to study diseases involving viral and bacterial pathogens and polymicrobial infections of the upper respiratory tract and the ear, such as otitis media. The value of the chinchilla as a model for human diseases prompted the sequencing of its genome in 2012 and the more recent development of the Chinchilla Research Resource Database (http://crrd.mcw.edu) to provide investigators with easy access to relevant datasets and software tools to enhance their research. The Chinchilla Research Resource Database contains a complete catalog of genes for chinchilla and, for comparative purposes, human. Chinchilla genes can be viewed in the context of their genomic scaffold positions using the JBrowse genome browser. In contrast to the corresponding records at NCBI, individual gene reports at CRRD include functional annotations for Disease, Gene Ontology (GO) Biological Process, GO Molecular Function, GO Cellular Component and Pathway assigned to chinchilla genes based on annotations from the corresponding human orthologs. Data can be retrieved via keyword and gene-specific searches. Lists of genes with similar functional attributes can be assembled by leveraging the hierarchical structure of the Disease, GO and Pathway vocabularies through the Ontology Search and Browser tool. Such lists can then be further analyzed for commonalities using the Gene Annotator (GA) Tool. All data in the Chinchilla Research Resource Database is freely accessible and downloadable via the CRRD FTP site or using the download functions available in the search and analysis tools. The Chinchilla Research Resource Database is a rich resource for researchers using, or considering the use of, chinchilla as a model for human disease. Database URL: http://crrd.mcw.edu PMID:27173523

  8. A scalable database model for multiparametric time series: a volcano observatory case study

    NASA Astrophysics Data System (ADS)

    Montalto, Placido; Aliotta, Marco; Cassisi, Carmelo; Prestifilippo, Michele; Cannata, Andrea

    2014-05-01

    The variables collected by a sensor network constitute a heterogeneous data source that needs to be properly organized in order to be used in research and geophysical monitoring. With the time series term we refer to a set of observations of a given phenomenon acquired sequentially in time. When the time intervals are equally spaced one speaks of period or sampling frequency. Our work describes in detail a possible methodology for storage and management of time series using a specific data structure. We designed a framework, hereinafter called TSDSystem (Time Series Database System), in order to acquire time series from different data sources and standardize them within a relational database. The operation of standardization provides the ability to perform operations, such as query and visualization, of many measures synchronizing them using a common time scale. The proposed architecture follows a multiple layer paradigm (Loaders layer, Database layer and Business Logic layer). Each layer is specialized in performing particular operations for the reorganization and archiving of data from different sources such as ASCII, Excel, ODBC (Open DataBase Connectivity), file accessible from the Internet (web pages, XML). In particular, the loader layer performs a security check of the working status of each running software through an heartbeat system, in order to automate the discovery of acquisition issues and other warning conditions. Although our system has to manage huge amounts of data, performance is guaranteed by using a smart partitioning table strategy, that keeps balanced the percentage of data stored in each database table. TSDSystem also contains modules for the visualization of acquired data, that provide the possibility to query different time series on a specified time range, or follow the realtime signal acquisition, according to a data access policy from the users.

  9. Using the Reactome Database

    PubMed Central

    Haw, Robin

    2012-01-01

    There is considerable interest in the bioinformatics community in creating pathway databases. The Reactome project (a collaboration between the Ontario Institute for Cancer Research, Cold Spring Harbor Laboratory, New York University Medical Center and the European Bioinformatics Institute) is one such pathway database and collects structured information on all the biological pathways and processes in the human. It is an expert-authored and peer-reviewed, curated collection of well-documented molecular reactions that span the gamut from simple intermediate metabolism to signaling pathways and complex cellular events. This information is supplemented with likely orthologous molecular reactions in mouse, rat, zebrafish, worm and other model organisms. This unit describes how to use the Reactome database to learn the steps of a biological pathway; navigate and browse through the Reactome database; identify the pathways in which a molecule of interest is involved; use the Pathway and Expression analysis tools to search the database for and visualize possible connections within user-supplied experimental data set and Reactome pathways; and the Species Comparison tool to compare human and model organism pathways. PMID:22700314

  10. Kriging-based generation of optimal databases as forward and inverse surrogate models

    NASA Astrophysics Data System (ADS)

    Bilicz, S.; Lambert, M.; Gyimóthy, Sz

    2010-07-01

    Numerical methods are used to simulate mathematical models for a wide range of engineering problems. The precision provided by such simulators is usually fine, but at the price of computational cost. In some applications this cost might be crucial. This leads us to consider cheap surrogate models in order to reduce the computation time still meeting the precision requirements. Among all available surrogate models, we deal herein with the generation of an 'optimal' database of pre-calculated results combined with a simple interpolator. A database generation approach is investigated which is intended to achieve an optimal sampling. Such databases can be used for the approximate solution of both forward and inverse problems. Their structure carries some meta-information about the involved physical problem. In the case of the inverse problem, an approach for predicting the uncertainty of the solution (due to the applied surrogate model and/or the uncertainty of the measured data) is presented. All methods are based on kriging—a stochastic tool for function approximation. Illustrative examples are drawn from eddy current non-destructive evaluation.

  11. Geodetic use of global digital terrain and crustal databases in gravity field modeling and interpretation

    NASA Astrophysics Data System (ADS)

    Tsoulis, D.

    2013-03-01

    The release of global digital databases for the description of the Earth's topography and the shape of the Earth's crust in terms of consistency and geometry initiates a new era in the interpretation and analysis of the observed gravity field of our planet. The permanent increase in resolution of these databases permits furthermore the identification of high frequency gravity field components, a feature that is of special interest in applications of local or regional scales. The derivation of topographic/isostatic gravity models is the tool which reveals the gravity content of terrain and crustal databases in the spectral domain. We review the significance of some current global digital models in the frame of this analysis by computing distinct spectral gravity quantities and compare them against the Kaula rule of the gravity signal decay and the recently released reference gravity model EGM2008. The different isostatic hypothesis that can be applied in the derivation of a topographic/isostatic model as well its dependency with the increasing harmonic degree is demonstrated and quantified in terms of geoid heights and gravity anomalies. It is shown that the two fundamental compensation mechanisms, namely Airy and Pratt, act complementary in terms of their compensation effect to the uncompensated topography spectrum. The Airy mechanism reduces the uncompensated topography in the longer and medium wavelength part of the spectrum (up to degree 400), while Pratt acts in a compensating manner only for the high to very high frequencies, from degree 100 and onwards.

  12. Leaf respiration (GlobResp) - global trait database supports Earth System Models

    SciTech Connect

    Wullschleger, Stan D.; Warren, Jeffrey; Thornton, Peter E.

    2015-03-20

    Here we detail how Atkin and his colleagues compiled a global database (GlobResp) that details rates of leaf dark respiration and associated traits from sites that span Arctic tundra to tropical forests. This compilation builds upon earlier research (Reich et al., 1998; Wright et al., 2006) and was supplemented by recent field campaigns and unpublished data.In keeping with other trait databases, GlobResp provides insights on how physiological traits, especially rates of dark respiration, vary as a function of environment and how that variation can be used to inform terrestrial biosphere models and land surface components of Earth System Models. Although an important component of plant and ecosystem carbon (C) budgets (Wythers et al., 2013), respiration has only limited representation in models. Seen through the eyes of a plant scientist, Atkin et al. (2015) give readers a unique perspective on the climatic controls on respiration, thermal acclimation and evolutionary adaptation of dark respiration, and insights into the covariation of respiration with other leaf traits. We find there is ample evidence that once large databases are compiled, like GlobResp, they can reveal new knowledge of plant function and provide a valuable resource for hypothesis testing and model development.

  13. Leaf respiration (GlobResp) - global trait database supports Earth System Models

    DOE PAGESBeta

    Wullschleger, Stan D.; Warren, Jeffrey; Thornton, Peter E.

    2015-03-20

    Here we detail how Atkin and his colleagues compiled a global database (GlobResp) that details rates of leaf dark respiration and associated traits from sites that span Arctic tundra to tropical forests. This compilation builds upon earlier research (Reich et al., 1998; Wright et al., 2006) and was supplemented by recent field campaigns and unpublished data.In keeping with other trait databases, GlobResp provides insights on how physiological traits, especially rates of dark respiration, vary as a function of environment and how that variation can be used to inform terrestrial biosphere models and land surface components of Earth System Models. Althoughmore » an important component of plant and ecosystem carbon (C) budgets (Wythers et al., 2013), respiration has only limited representation in models. Seen through the eyes of a plant scientist, Atkin et al. (2015) give readers a unique perspective on the climatic controls on respiration, thermal acclimation and evolutionary adaptation of dark respiration, and insights into the covariation of respiration with other leaf traits. We find there is ample evidence that once large databases are compiled, like GlobResp, they can reveal new knowledge of plant function and provide a valuable resource for hypothesis testing and model development.« less

  14. Modeling Personnel Turnover in the Parametric Organization

    NASA Technical Reports Server (NTRS)

    Dean, Edwin B.

    1991-01-01

    A primary issue in organizing a new parametric cost analysis function is to determine the skill mix and number of personnel required. The skill mix can be obtained by a functional decomposition of the tasks required within the organization and a matrixed correlation with educational or experience backgrounds. The number of personnel is a function of the skills required to cover all tasks, personnel skill background and cross training, the intensity of the workload for each task, migration through various tasks by personnel along a career path, personnel hiring limitations imposed by management and the applicant marketplace, personnel training limitations imposed by management and personnel capability, and the rate at which personnel leave the organization for whatever reason. Faced with the task of relating all of these organizational facets in order to grow a parametric cost analysis (PCA) organization from scratch, it was decided that a dynamic model was required in order to account for the obvious dynamics of the forming organization. The challenge was to create such a simple model which would be credible during all phases of organizational development. The model development process was broken down into the activities of determining the tasks required for PCA, determining the skills required for each PCA task, determining the skills available in the applicant marketplace, determining the structure of the dynamic model, implementing the dynamic model, and testing the dynamic model.

  15. Data-based stochastic subgrid-scale parametrization: an approach using cluster-weighted modelling.

    PubMed

    Kwasniok, Frank

    2012-03-13

    A new approach for data-based stochastic parametrization of unresolved scales and processes in numerical weather and climate prediction models is introduced. The subgrid-scale model is conditional on the state of the resolved scales, consisting of a collection of local models. A clustering algorithm in the space of the resolved variables is combined with statistical modelling of the impact of the unresolved variables. The clusters and the parameters of the associated subgrid models are estimated simultaneously from data. The method is implemented and explored in the framework of the Lorenz '96 model using discrete Markov processes as local statistical models. Performance of the cluster-weighted Markov chain scheme is investigated for long-term simulations as well as ensemble prediction. It clearly outperforms simple parametrization schemes and compares favourably with another recently proposed subgrid modelling scheme also based on conditional Markov chains.

  16. Condensing Organic Aerosols in a Microphysical Model

    NASA Astrophysics Data System (ADS)

    Gao, Y.; Tsigaridis, K.; Bauer, S.

    2015-12-01

    The condensation of organic aerosols is represented in a newly developed box-model scheme, where its effect on the growth and composition of particles are examined. We implemented the volatility-basis set (VBS) framework into the aerosol mixing state resolving microphysical scheme Multiconfiguration Aerosol TRacker of mIXing state (MATRIX). This new scheme is unique and advances the representation of organic aerosols in models in that, contrary to the traditional treatment of organic aerosols as non-volatile in most climate models and in the original version of MATRIX, this new scheme treats them as semi-volatile. Such treatment is important because low-volatility organics contribute significantly to the growth of particles. The new scheme includes several classes of semi-volatile organic compounds from the VBS framework that can partition among aerosol populations in MATRIX, thus representing the growth of particles via condensation of low volatility organic vapors. Results from test cases representing Mexico City and a Finish forrest condistions show good representation of the time evolutions of concentration for VBS species in the gas phase and in the condensed particulate phase. Emitted semi-volatile primary organic aerosols evaporate almost completely in the high volatile range, and they condense more efficiently in the low volatility range.

  17. Modelling nitrous oxide emissions from organic soils in Europe

    NASA Astrophysics Data System (ADS)

    Leppelt, Thomas; Dechow, Rene; Gebbert, Sören; Freibauer, Annette

    2013-04-01

    The greenhouse gas emission potential of peatland ecosystems are mandatory for a complete annual emission budget in Europe. The GHG-Europe project aims to improve the modelling capabilities for greenhouse gases, e.g., nitrous oxide. The heterogeneous and event driven fluxes of nitrous oxide are challenging to model on European scale, especially regarding the upscaling purpose and certain parameter estimations. Due to these challenges adequate techniques are needed to create a robust empirical model. Therefore a literature study of nitrous oxide fluxes from organic soils has been carried out. This database contains flux data from boreal and temperate climate zones and covers the different land use categories: cropland, grassland, forest, natural and peat extraction sites. Especially managed crop- and grassland sites feature high emission potential. Generally nitrous oxide emissions increases significantly with deep drainage and intensive application of nitrogen fertilisation. Whereas natural peatland sites with a near surface groundwater table can act as nitrous oxide sink. An empirical fuzzy logic model has been applied to predict annual nitrous oxide emissions from organic soils. The calibration results in two separate models with best model performances for bogs and fens, respectively. The derived parameter combinations of these models contain mean groundwater table, nitrogen fertilisation, annual precipitation, air temperature, carbon content and pH value. Influences of the calibrated parameters on nitrous oxide fluxes are verified by several studies in literature. The extrapolation potential has been tested by an implemented cross validation. Furthermore the parameter ranges of the calibrated models are compared to occurring values on European scale. This avoid unknown systematic errors for the regionalisation purpose. Additionally a sensitivity analysis specify the model behaviour for each alternating parameter. The upscaling process for European peatland

  18. Database of the United States Coal Pellet Collection of the U.S. Geological Survey Organic Petrology Laboratory

    USGS Publications Warehouse

    Deems, Nikolaus J.; Hackley, Paul C.

    2012-01-01

    The Organic Petrology Laboratory (OPL) of the U.S. Geological Survey (USGS) Eastern Energy Resources Science Center in Reston, Virginia, contains several thousand processed coal sample materials that were loosely organized in laboratory drawers for the past several decades. The majority of these were prepared as 1-inch-diameter particulate coal pellets (more than 6,000 pellets; one sample usually was prepared as two pellets, although some samples were prepared in as many as four pellets), which were polished and used in reflected light petrographic studies. These samples represent the work of many scientists from the 1970s to the present, most notably Ron Stanton, who managed the OPL until 2001 (see Warwick and Ruppert, 2005, for a comprehensive bibliography of Ron Stanton's work). The purpose of the project described herein was to organize and catalog the U.S. part of the petrographic sample collection into a comprehensive database (available with this report as a Microsoft Excel file) and to compile and list published studies associated with the various sample sets. Through this work, the extent of the collection is publicly documented as a resource and sample library available to other scientists and researchers working in U.S. coal basins previously studied by organic petrologists affiliated with the USGS. Other researchers may obtain samples in the OPL collection on loan at the discretion of the USGS authors listed in this report and its associated Web page.

  19. Toxicity of halogenated organic compounds. (Latest citations from the NTIS bibliographic database). Published Search

    SciTech Connect

    Not Available

    1993-09-01

    The bibliography contains citations concerning health and environmental effects of halogenated organic compounds. Topics include laboratory and field investigations regarding bioaccumulation and concentration, metabolic aspects, and specific site studies in industrial and commercial operations. Pesticides, solvents, and a variety of industrial compounds are discussed. (Contains 250 citations and includes a subject term index and title list.)

  20. Validation of a Database of Formation Enthalpies and of Mid-Level Model Chemistries.

    PubMed

    Simmie, J M; Sheahan, J N

    2016-09-22

    In order to test new procedures for the calculation of basic molecular properties, a properly validated database and computational method appropriate to the range of species at hand is essential. Here formation enthalpies of chemical species CmHnNpOq from their constituent atoms are computed by midlevel composite model chemistries in order to check the contents of the best established and most accurate database, ATcT. Once discrepancies are identified alternative independent procedures and/or higher level model chemistries, which include CCSDT(Q) calculations, are employed to resolve the problems. Shortcomings of the midlevel methods used are signaled where these occur. In addition a more visual statistical analysis than is usual is presented which highlights the outliers and identifies the bias of each method together with associated error bars and the 95% limits of agreement and its error bars. PMID:27583807

  1. S-World: A high resolution global soil database for simulation modelling (Invited)

    NASA Astrophysics Data System (ADS)

    Stoorvogel, J. J.

    2013-12-01

    There is an increasing call for high resolution soil information at the global level. A good example for such a call is the Global Gridded Crop Model Intercomparison carried out within AgMIP. While local studies can make use of surveying techniques to collect additional techniques this is practically impossible at the global level. It is therefore important to rely on legacy data like the Harmonized World Soil Database. Several efforts do exist that aim at the development of global gridded soil property databases. These estimates of the variation of soil properties can be used to assess e.g., global soil carbon stocks. However, they do not allow for simulation runs with e.g., crop growth simulation models as these models require a description of the entire pedon rather than a few soil properties. This study provides the required quantitative description of pedons at a 1 km resolution for simulation modelling. It uses the Harmonized World Soil Database (HWSD) for the spatial distribution of soil types, the ISRIC-WISE soil profile database to derive information on soil properties per soil type, and a range of co-variables on topography, climate, and land cover to further disaggregate the available data. The methodology aims to take stock of these available data. The soil database is developed in five main steps. Step 1: All 148 soil types are ordered on the basis of their expected topographic position using e.g., drainage, salinization, and pedogenesis. Using the topographic ordering and combining the HWSD with a digital elevation model allows for the spatial disaggregation of the composite soil units. This results in a new soil map with homogeneous soil units. Step 2: The ranges of major soil properties for the topsoil and subsoil of each of the 148 soil types are derived from the ISRIC-WISE soil profile database. Step 3: A model of soil formation is developed that focuses on the basic conceptual question where we are within the range of a particular soil property

  2. Conceptual database modeling: a method for enabling end users (radiologists) to understand and develop their information management applications.

    PubMed

    Hawkins, H; Young, S K; Hubert, K C; Hallock, P

    2001-06-01

    As medical technology advances at a rapid pace, clinicians become further and further removed from the design of their own technological tools. This is particularly evident with information management. For radiologists, clinical histories, patient reports, and other pertinent information require sophisticated tools for data handling. However, as databases grow more powerful and sophisticated, systems require the expertise of programmers and information technology personnel. The radiologist, the clinician end-user, must maintain involvement in the development of system tools to insure effective information management. Conceptual database modeling is a design method that serves to bridge the gap between the technological aspects of information management and its clinical applications. Conceptual database modeling involves developing information systems in simple language so that anyone can have input into the overall design. This presentation describes conceptual database modeling, using object role modeling, as a means by which end-users (clinicians) may participate in database development.

  3. Object-oriented urban 3D spatial data model organization method

    NASA Astrophysics Data System (ADS)

    Li, Jing-wen; Li, Wen-qing; Lv, Nan; Su, Tao

    2015-12-01

    This paper combined the 3d data model with object-oriented organization method, put forward the model of 3d data based on object-oriented method, implemented the city 3d model to quickly build logical semantic expression and model, solved the city 3d spatial information representation problem of the same location with multiple property and the same property with multiple locations, designed the space object structure of point, line, polygon, body for city of 3d spatial database, and provided a new thought and method for the city 3d GIS model and organization management.

  4. Sediment-Hosted Copper Deposits of the World: Deposit Models and Database

    USGS Publications Warehouse

    Cox, Dennis P.; Lindsey, David A.; Singer, Donald A.; Diggles, Michael F.

    2003-01-01

    Introduction This publication contains four descriptive models and four grade-tonnage models for sediment hosted copper deposits. Descriptive models are useful in exploration planning and resource assessment because they enable the user to identify deposits in the field and to identify areas on geologic and geophysical maps where deposits could occur. Grade and tonnage models are used in resource assessment to predict the likelihood of different combinations of grades and tonnages that could occur in undiscovered deposits in a specific area. They are also useful in exploration in deciding what deposit types meet the economic objectives of the exploration company. The models in this report supersede the sediment-hosted copper models in USGS Bulletin 1693 (Cox, 1986, and Mosier and others, 1986) and are subdivided into a general type and three subtypes. The general model is useful in classifying deposits whose features are obscured by metamorphism or are otherwise poorly described, and for assessing regions in which the geologic environments are poorly understood. The three subtypes are based on differences in deposit form and environments of deposition. These differences are described under subtypes in the general model. Deposit models are based on the descriptions of geologic environments and physical characteristics, and on metal grades and tonnages of many individual deposits. Data used in this study are presented in a database representing 785 deposits in nine continents. This database was derived partly from data published by Kirkham and others (1994) and from new information in recent publications. To facilitate the construction of grade and tonnage models, the information, presented by Kirkham in disaggregated form, was brought together to provide a single grade and a single tonnage for each deposit. Throughout the report individual deposits are defined as being more than 2,000 meters from the nearest adjacent deposit. The deposit models are presented here as

  5. The origin and evolution of model organisms

    NASA Technical Reports Server (NTRS)

    Hedges, S. Blair

    2002-01-01

    The phylogeny and timescale of life are becoming better understood as the analysis of genomic data from model organisms continues to grow. As a result, discoveries are being made about the early history of life and the origin and development of complex multicellular life. This emerging comparative framework and the emphasis on historical patterns is helping to bridge barriers among organism-based research communities.

  6. Open systems and databases

    SciTech Connect

    Martire, G.S. ); Nuttall, D.J.H. )

    1993-05-01

    This paper is part of a series of papers invited by the IEEE POWER CONTROL CENTER WORKING GROUP concerning the changing designs of modern control centers. Papers invited by the Working Group discuss the following issues: Benefits of Openness, Criteria for Evaluating Open EMS Systems, Hardware Design, Configuration Management, Security, Project Management, Databases, SCADA, Inter- and Intra-System Communications and Man-Machine Interfaces,'' The goal of this paper is to provide an introduction to the issues pertaining to Open Systems and Databases.'' The intent is to assist understanding of some of the underlying factors that effect choices that must be made when selecting a database system for use in a control room environment. This paper describes and compares the major database information models which are in common use for database systems and provides an overview of SQL. A case for the control center community to follow the workings of the non-formal standards bodies is presented along with possible uses and the benefits of commercially available databases within the control center. The reasons behind the emergence of industry supported standards organizations such as the Open Software Foundation (OSF) and SQL Access are presented.

  7. Research proceedings on amphibian model organisms.

    PubMed

    Liu, Lu-Sha; Zhao, Lan-Ying; Wang, Shou-Hong; Jiang, Jian-Ping

    2016-07-18

    Model organisms have long been important in biology and medicine due to their specific characteristics. Amphibians, especially Xenopus, play key roles in answering fundamental questions on developmental biology, regeneration, genetics, and toxicology due to their large and abundant eggs, as well as their versatile embryos, which can be readily manipulated and developed in vivo. Furthermore, amphibians have also proven to be of considerable benefit in human disease research due to their conserved cellular developmental and genomic organization. This review gives a brief introduction on the progress and limitations of these animal models in biology and human disease research, and discusses the potential and challenge of Microhyla fissipes as a new model organism. PMID:27469255

  8. Research proceedings on amphibian model organisms

    PubMed Central

    LIU, Lu-Sha; ZHAO, Lan-Ying; WANG, Shou-Hong; JIANG, Jian-Ping

    2016-01-01

    Model organisms have long been important in biology and medicine due to their specific characteristics. Amphibians, especially Xenopus, play key roles in answering fundamental questions on developmental biology, regeneration, genetics, and toxicology due to their large and abundant eggs, as well as their versatile embryos, which can be readily manipulated and developed in vivo. Furthermore, amphibians have also proven to be of considerable benefit in human disease research due to their conserved cellular developmental and genomic organization. This review gives a brief introduction on the progress and limitations of these animal models in biology and human disease research, and discusses the potential and challenge of Microhyla fissipes as a new model organism. PMID:27469255

  9. Putting "Organizations" into an Organization Theory Course: A Hybrid CAO Model for Teaching Organization Theory

    ERIC Educational Resources Information Center

    Hannah, David R.; Venkatachary, Ranga

    2010-01-01

    In this article, the authors present a retrospective analysis of an instructor's multiyear redesign of a course on organization theory into what is called a hybrid Classroom-as-Organization model. It is suggested that this new course design served to apprentice students to function in quasi-real organizational structures. The authors further argue…

  10. Very fast road database verification using textured 3D city models obtained from airborne imagery

    NASA Astrophysics Data System (ADS)

    Bulatov, Dimitri; Ziems, Marcel; Rottensteiner, Franz; Pohl, Melanie

    2014-10-01

    Road databases are known to be an important part of any geodata infrastructure, e.g. as the basis for urban planning or emergency services. Updating road databases for crisis events must be performed quickly and with the highest possible degree of automation. We present a semi-automatic algorithm for road verification using textured 3D city models, starting from aerial or even UAV-images. This algorithm contains two processes, which exchange input and output, but basically run independently from each other. These processes are textured urban terrain reconstruction and road verification. The first process contains a dense photogrammetric reconstruction of 3D geometry of the scene using depth maps. The second process is our core procedure, since it contains various methods for road verification. Each method represents a unique road model and a specific strategy, and thus is able to deal with a specific type of roads. Each method is designed to provide two probability distributions, where the first describes the state of a road object (correct, incorrect), and the second describes the state of its underlying road model (applicable, not applicable). Based on the Dempster-Shafer Theory, both distributions are mapped to a single distribution that refers to three states: correct, incorrect, and unknown. With respect to the interaction of both processes, the normalized elevation map and the digital orthophoto generated during 3D reconstruction are the necessary input - together with initial road database entries - for the road verification process. If the entries of the database are too obsolete or not available at all, sensor data evaluation enables classification of the road pixels of the elevation map followed by road map extraction by means of vectorization and filtering of the geometrically and topologically inconsistent objects. Depending on the time issue and availability of a geo-database for buildings, the urban terrain reconstruction procedure has semantic models

  11. Polymer models of chromosome (re)organization

    NASA Astrophysics Data System (ADS)

    Mirny, Leonid

    Chromosome Conformation Capture technique (Hi-C) provides comprehensive information about frequencies of spatial interactions between genomic loci. Inferring 3D organization of chromosomes from these data is a challenging biophysical problem. We develop a top-down approach to biophysical modeling of chromosomes. Starting with a minimal set of biologically motivated interactions we build ensembles of polymer conformations that can reproduce major features observed in Hi-C experiments. I will present our work on modeling organization of human metaphase and interphase chromosomes. Our works suggests that active processes of loop extrusion can be a universal mechanism responsible for formation of domains in interphase and chromosome compaction in metaphase.

  12. Developmental database for phenology models: related insect and mite species have similar thermal requirements.

    PubMed

    Jarosík, Vojtech; Honek, Alois; Magarey, Roger D; Skuhrovec, Jirí

    2011-12-01

    Two values of thermal requirements, the lower developmental threshold (LDT), that is, the temperature at which development ceases, and the sum of effective temperatures, that is, day degrees above the LDT control the development of ectotherms and are used in phenology models to predict time at which the development of individual stages of a species will be completed. To assist in the rapid development of phenology models, we merged a previously published database of thermal requirements for insects, gathered by online search in CAB Abstracts, with independently collected data for insects and mites from original studies. The merged database comprises developmental times at various constant temperatures on 1,054 insect and mite species, many of them in several populations, mostly pests and their natural enemies, from all over the world. We show that closely related species share similar thermal requirements and therefore, for a species with unknown thermal requirements, the value of LDT and sum of effective temperatures of its most related species from the database can be used. PMID:22299347

  13. Thermodynamic Model and Database for Sulfides Dissolved in Molten Oxide Slags

    NASA Astrophysics Data System (ADS)

    Kang, Youn-Bae; Pelton, Arthur D.

    2009-12-01

    A thermodynamic model has been developed in the framework of the modified quasichemical model in the quadruplet approximation to permit the calculation of solubilities of various gaseous species (sulfide, sulfate, nitride, carbide, water, etc.) in molten slags. The model calculates the solubilities solely from knowledge of the thermodynamic activities of the component oxides and the Gibbs energies of the pure liquid components (oxides, sulfides, sulfates, etc.). In the current article, it is shown that solubilities of sulfur as sulfide in Al2O3-CaO-FeO-Fe2O3-MgO-MnO-SiO2-TiO2-Ti2O3 multicomponent slags, which are predicted from the current model with no adjustable model parameters, are in good agreement with all available experimental data. The article also provides a thorough review of experimental sulfide capacity data for this system. The model applies at all compositions from pure oxides to pure sulfides and from basic to acidic slags. By coupling this database with other evaluated databases, such as those for molten metal and gaseous phases, and with general software for Gibbs energy minimization, practically important slag/metal/gas/solid equilibria can be computed such as S-distribution ratios.

  14. ModBase, a database of annotated comparative protein structure models, and associated resources.

    PubMed

    Pieper, Ursula; Webb, Benjamin M; Barkan, David T; Schneidman-Duhovny, Dina; Schlessinger, Avner; Braberg, Hannes; Yang, Zheng; Meng, Elaine C; Pettersen, Eric F; Huang, Conrad C; Datta, Ruchira S; Sampathkumar, Parthasarathy; Madhusudhan, Mallur S; Sjölander, Kimmen; Ferrin, Thomas E; Burley, Stephen K; Sali, Andrej

    2011-01-01

    ModBase (http://salilab.org/modbase) is a database of annotated comparative protein structure models. The models are calculated by ModPipe, an automated modeling pipeline that relies primarily on Modeller for fold assignment, sequence-structure alignment, model building and model assessment (http://salilab.org/modeller/). ModBase currently contains 10,355,444 reliable models for domains in 2,421,920 unique protein sequences. ModBase allows users to update comparative models on demand, and request modeling of additional sequences through an interface to the ModWeb modeling server (http://salilab.org/modweb). ModBase models are available through the ModBase interface as well as the Protein Model Portal (http://www.proteinmodelportal.org/). Recently developed associated resources include the SALIGN server for multiple sequence and structure alignment (http://salilab.org/salign), the ModEval server for predicting the accuracy of protein structure models (http://salilab.org/modeval), the PCSS server for predicting which peptides bind to a given protein (http://salilab.org/pcss) and the FoXS server for calculating and fitting Small Angle X-ray Scattering profiles (http://salilab.org/foxs). PMID:21097780

  15. Seismic hazard assessment for Myanmar: Earthquake model database, ground-motion scenarios, and probabilistic assessments

    NASA Astrophysics Data System (ADS)

    Chan, C. H.; Wang, Y.; Thant, M.; Maung Maung, P.; Sieh, K.

    2015-12-01

    We have constructed an earthquake and fault database, conducted a series of ground-shaking scenarios, and proposed seismic hazard maps for all of Myanmar and hazard curves for selected cities. Our earthquake database integrates the ISC, ISC-GEM and global ANSS Comprehensive Catalogues, and includes harmonized magnitude scales without duplicate events. Our active fault database includes active fault data from previous studies. Using the parameters from these updated databases (i.e., the Gutenberg-Richter relationship, slip rate, maximum magnitude and the elapse time of last events), we have determined the earthquake recurrence models of seismogenic sources. To evaluate the ground shaking behaviours in different tectonic regimes, we conducted a series of tests by matching the modelled ground motions to the felt intensities of earthquakes. Through the case of the 1975 Bagan earthquake, we determined that Atkinson and Moore's (2003) scenario using the ground motion prediction equations (GMPEs) fits the behaviours of the subduction events best. Also, the 2011 Tarlay and 2012 Thabeikkyin events suggested the GMPEs of Akkar and Cagnan (2010) fit crustal earthquakes best. We thus incorporated the best-fitting GMPEs and site conditions based on Vs30 (the average shear-velocity down to 30 m depth) from analysis of topographic slope and microtremor array measurements to assess seismic hazard. The hazard is highest in regions close to the Sagaing Fault and along the Western Coast of Myanmar as seismic sources there have earthquakes occur at short intervals and/or last events occurred a long time ago. The hazard curves for the cities of Bago, Mandalay, Sagaing, Taungoo and Yangon show higher hazards for sites close to an active fault or with a low Vs30, e.g., the downtown of Sagaing and Shwemawdaw Pagoda in Bago.

  16. Thermodynamic modeling for organic solid precipitation

    SciTech Connect

    Chung, T.H.

    1992-12-01

    A generalized predictive model which is based on thermodynamic principle for solid-liquid phase equilibrium has been developed for organic solid precipitation. The model takes into account the effects of temperature, composition, and activity coefficient on the solubility of wax and asphaltenes in organic solutions. The solid-liquid equilibrium K-value is expressed as a function of the heat of melting, melting point temperature, solubility parameter, and the molar volume of each component in the solution. All these parameters have been correlated with molecular weight. Thus, the model can be applied to crude oil systems. The model has been tested with experimental data for wax formation and asphaltene precipitation. The predicted wax appearance temperature is very close to the measured temperature. The model not only can match the measured asphaltene solubility data but also can be used to predict the solubility of asphaltene in organic solvents or crude oils. The model assumes that asphaltenes are dissolved in oil in a true liquid state, not in colloidal suspension, and the precipitation-dissolution process is reversible by changing thermodynamic conditions. The model is thermodynamically consistent and has no ambiguous assumptions.

  17. Solubility Database

    National Institute of Standards and Technology Data Gateway

    SRD 106 IUPAC-NIST Solubility Database (Web, free access)   These solubilities are compiled from 18 volumes (Click here for List) of the International Union for Pure and Applied Chemistry(IUPAC)-NIST Solubility Data Series. The database includes liquid-liquid, solid-liquid, and gas-liquid systems. Typical solvents and solutes include water, seawater, heavy water, inorganic compounds, and a variety of organic compounds such as hydrocarbons, halogenated hydrocarbons, alcohols, acids, esters and nitrogen compounds. There are over 67,500 solubility measurements and over 1800 references.

  18. Finding mouse models of human lymphomas and leukemia's using the Jackson laboratory mouse tumor biology database.

    PubMed

    Begley, Dale A; Sundberg, John P; Krupke, Debra M; Neuhauser, Steven B; Bult, Carol J; Eppig, Janan T; Morse, Herbert C; Ward, Jerrold M

    2015-12-01

    Many mouse models have been created to study hematopoietic cancer types. There are over thirty hematopoietic tumor types and subtypes, both human and mouse, with various origins, characteristics and clinical prognoses. Determining the specific type of hematopoietic lesion produced in a mouse model and identifying mouse models that correspond to the human subtypes of these lesions has been a continuing challenge for the scientific community. The Mouse Tumor Biology Database (MTB; http://tumor.informatics.jax.org) is designed to facilitate use of mouse models of human cancer by providing detailed histopathologic and molecular information on lymphoma subtypes, including expertly annotated, on line, whole slide scans, and providing a repository for storing information on and querying these data for specific lymphoma models. PMID:26302176

  19. Coupling ensemble weather predictions based on TIGGE database with Grid-Xinanjiang model for flood forecast

    NASA Astrophysics Data System (ADS)

    Bao, H.-J.; Zhao, L.-N.; He, Y.; Li, Z.-J.; Wetterhall, F.; Cloke, H. L.; Pappenberger, F.; Manful, D.

    2011-02-01

    The incorporation of numerical weather predictions (NWP) into a flood forecasting system can increase forecast lead times from a few hours to a few days. A single NWP forecast from a single forecast centre, however, is insufficient as it involves considerable non-predictable uncertainties and lead to a high number of false alarms. The availability of global ensemble numerical weather prediction systems through the THORPEX Interactive Grand Global Ensemble' (TIGGE) offers a new opportunity for flood forecast. The Grid-Xinanjiang distributed hydrological model, which is based on the Xinanjiang model theory and the topographical information of each grid cell extracted from the Digital Elevation Model (DEM), is coupled with ensemble weather predictions based on the TIGGE database (CMC, CMA, ECWMF, UKMO, NCEP) for flood forecast. This paper presents a case study using the coupled flood forecasting model on the Xixian catchment (a drainage area of 8826 km2) located in Henan province, China. A probabilistic discharge is provided as the end product of flood forecast. Results show that the association of the Grid-Xinanjiang model and the TIGGE database gives a promising tool for an early warning of flood events several days ahead.

  20. Assessment of the health effects of chemicals in humans: II. Construction of an adverse effects database for QSAR modeling.

    PubMed

    Matthews, Edwin J; Kruhlak, Naomi L; Weaver, James L; Benz, R Daniel; Contrera, Joseph F

    2004-12-01

    The FDA's Spontaneous Reporting System (SRS) database contains over 1.5 million adverse drug reaction (ADR) reports for 8620 drugs/biologics that are listed for 1191 Coding Symbols for Thesaurus of Adverse Reaction (COSTAR) terms of adverse effects. We have linked the trade names of the drugs to 1861 generic names and retrieved molecular structures for each chemical to obtain a set of 1515 organic chemicals that are suitable for modeling with commercially available QSAR software packages. ADR report data for 631 of these compounds were extracted and pooled for the first five years that each drug was marketed. Patient exposure was estimated during this period using pharmaceutical shipping units obtained from IMS Health. Significant drug effects were identified using a Reporting Index (RI), where RI = (# ADR reports / # shipping units) x 1,000,000. MCASE/MC4PC software was used to identify the optimal conditions for defining a significant adverse effect finding. Results suggest that a significant effect in our database is characterized by > or = 4 ADR reports and > or = 20,000 shipping units during five years of marketing, and an RI > or = 4.0. Furthermore, for a test chemical to be evaluated as active it must contain a statistically significant molecular structural alert, called a decision alert, in two or more toxicologically related endpoints. We also report the use of a composite module, which pools observations from two or more toxicologically related COSTAR term endpoints to provide signal enhancement for detecting adverse effects. PMID:16472241

  1. SMIfp (SMILES fingerprint) chemical space for virtual screening and visualization of large databases of organic molecules.

    PubMed

    Schwartz, Julian; Awale, Mahendra; Reymond, Jean-Louis

    2013-08-26

    SMIfp (SMILES fingerprint) is defined here as a scalar fingerprint describing organic molecules by counting the occurrences of 34 different symbols in their SMILES strings, which creates a 34-dimensional chemical space. Ligand-based virtual screening using the city-block distance CBD(SMIfp) as similarity measure provides good AUC values and enrichment factors for recovering series of actives from the directory of useful decoys (DUD-E) and from ZINC. DrugBank, ChEMBL, ZINC, PubChem, GDB-11, GDB-13, and GDB-17 can be searched by CBD(SMIfp) using an online SMIfp-browser at www.gdb.unibe.ch. Visualization of the SMIfp chemical space was performed by principal component analysis and color-coded maps of the (PC1, PC2)-planes, with interactive access to the molecules enabled by the Java application SMIfp-MAPPLET available from www.gdb.unibe.ch. These maps spread molecules according to their fraction of aromatic atoms, size and polarity. SMIfp provides a new and relevant entry to explore the small molecule chemical space.

  2. GIS database model for development of mining information system for the Skarn/Porphyry type ore deposit

    NASA Astrophysics Data System (ADS)

    Roh, T.; Choi, Y.; Park, H.

    2009-12-01

    This study presents a prototype of GIS database model for development of mining information system for the Skarn/Porphyry type ore deposit. Database table was established for the analysis of collected datum from mining activity and geological investigation of mine development. Also structure and property of geological/mining information elements composing each table were defined and specified. For each mine, mine shaft, line and point, independent ID code were assigened. Database is also designed to keep the graphic data of Stereophotogrammetry from mining working face and of geophysical and boring investigation. After combining existing mine map and digital elevation map, sample data was inputed to the database. Finally, database system model that can be used for additional development of mining information system was constructed in this study.

  3. Assessment of cloud cover in climate models and reanalysis databases with ISCCP over the Mediterranean region

    NASA Astrophysics Data System (ADS)

    Enriquez, Aaron; Calbo, Josep; Gonzalez, Josep-Abel

    2013-04-01

    Clouds are an important regulator of climate due to their influence on the water balance of the atmosphere and their interaction with solar and infrared radiation. At any time, clouds cover a great percentage of the Earth's surface but their distribution is very irregular along time and space, which makes the evaluation of their influence on climate a difficult task. At present there are few studies related to cloud cover comparing current climate models with observational data. In this study, the database of monthly cloud cover provided by the International Satellite Cloud Climatology Project (ISCCP) has been chosen as a reference against which we compare the output of CMIP5 climate models and reanalysis databases, on the domain South-Europe-Mediterranean (SEM) established by the Intergovernmental Panel on Climate Change (IPCC) [1]. The study covers the period between 1984 and 2009, and the performance of cloud cover estimations for seasons has also been studied. To quantify the agreement between the databases we use two types of statistics: bias and SkillScore, which is based on the probability density functions (PDFs) of the databases [2]. We also use Taylor diagrams to visualize the statistics. Results indicate that there are areas where the models accurately describe what it is observed by ISCCP, for some periods of the year (e.g. Northern Africa, for autumn), compared to other areas and periods for which the agreement is lower (Iberian Peninsula in winter and the Black Sea for the summer months). However these differences should be attributed not only to the limitations of climate models, but possibly also to the data provided by ISCCP. References [1] Intergovernmental Panel on Climate Change (2007) Fourth Assessment Report: Climate Change 2007: Working Group I Report: The Physical Science Basis. [2] Ranking the AR4 climate models over the Murray Darling Basin using simulated maximum temperature, minimum temperature and precipitation. Int J Climatol 28

  4. Improving Quality and Quantity of Contributions: Two Models for Promoting Knowledge Exchange with Shared Databases

    ERIC Educational Resources Information Center

    Cress, U.; Barquero, B.; Schwan, S.; Hesse, F. W.

    2007-01-01

    Shared databases are used for knowledge exchange in groups. Whether a person is willing to contribute knowledge to a shared database presents a social dilemma: Each group member saves time and energy by not contributing any information to the database and by using the database only to retrieve information which was contributed by others. But if…

  5. Biophysical Modeling of Respiratory Organ Motion

    NASA Astrophysics Data System (ADS)

    Werner, René

    Methods to estimate respiratory organ motion can be divided into two groups: biophysical modeling and image registration. In image registration, motion fields are directly extracted from 4D ({D}+{t}) image sequences, often without concerning knowledge about anatomy and physiology in detail. In contrast, biophysical approaches aim at identification of anatomical and physiological aspects of breathing dynamics that are to be modeled. In the context of radiation therapy, biophysical modeling of respiratory organ motion commonly refers to the framework of continuum mechanics and elasticity theory, respectively. Underlying ideas and corresponding boundary value problems of those approaches are described in this chapter, along with a brief comparison to image registration-based motion field estimation.

  6. Object-Oriented Database for Managing Building Modeling Components and Metadata: Preprint

    SciTech Connect

    Long, N.; Fleming, K.; Brackney, L.

    2011-12-01

    Building simulation enables users to explore and evaluate multiple building designs. When tools for optimization, parametrics, and uncertainty analysis are combined with analysis engines, the sheer number of discrete simulation datasets makes it difficult to keep track of the inputs. The integrity of the input data is critical to designers, engineers, and researchers for code compliance, validation, and building commissioning long after the simulations are finished. This paper discusses an application that stores inputs needed for building energy modeling in a searchable, indexable, flexible, and scalable database to help address the problem of managing simulation input data.

  7. BRAGI: linking and visualization of database information in a 3D viewer and modeling tool.

    PubMed

    Reichelt, Joachim; Dieterich, Guido; Kvesic, Marsel; Schomburg, Dietmar; Heinz, Dirk W

    2005-04-01

    BRAGI is a well-established package for viewing and modeling of three-dimensional (3D) structures of biological macromolecules. A new version of BRAGI has been developed that is supported on Windows, Linux and SGI. The user interface has been rewritten to give the standard 'look and feel' of the chosen operating system and to provide a more intuitive, easier usage. A large number of new features have been added. Information from public databases such as SWISS-PROT, InterPro, DALI and OMIM can be displayed in the 3D viewer. Structures can be searched for homologous sequences using the NCBI BLAST server.

  8. Conceptual Model Formalization in a Semantic Interoperability Service Framework: Transforming Relational Database Schemas to OWL.

    PubMed

    Bravo, Carlos; Suarez, Carlos; González, Carolina; López, Diego; Blobel, Bernd

    2014-01-01

    Healthcare information is distributed through multiple heterogeneous and autonomous systems. Access to, and sharing of, distributed information sources are a challenging task. To contribute to meeting this challenge, this paper presents a formal, complete and semi-automatic transformation service from Relational Databases to Web Ontology Language. The proposed service makes use of an algorithm that allows to transform several data models of different domains by deploying mainly inheritance rules. The paper emphasizes the relevance of integrating the proposed approach into an ontology-based interoperability service to achieve semantic interoperability.

  9. Conception and development of a bibliographic database of blood nutrient fluxes across organs and tissues in ruminants: data gathering and management prior to meta-analysis.

    PubMed

    Vernet, Jean; Ortigues-Marty, Isabelle

    2006-01-01

    In the organism, nutrient exchanges among tissues and organs are subject to numerous sources of physiological or nutritional variation, and the contribution of individual factors needs to be quantified before establishing general response laws. To achieve this, meta-analysis of data from publications is a useful tool. The objective of this work was to develop a bibliographic database of nutrient fluxes across organs and tissues of ruminant animals (Flora) under Access using the Merise method. The most important criteria for Flora were the ease to relate the various information, the exhaustivity and the accuracy of the data input, a complete description of the diets, taking into account the methods of the methodological procedures of measurement and analysis of blood nutrients and the traceability of the information. The conceptual data model was built in 6 parts. The first part describes the authors and source of publication, and the person in charge of data input. It clearly separates and identifies the experiments, the groups of animals and the treatments within a publication. The second part is concerned with feeds, diets and their chemical composition and nutritional value. The third and fourth parts describe the infusion of any substrates and the methods employed, respectively. The fifth part is devoted to the results of blood flows and nutrient fluxes. The sixth part gathers miscellaneous experimental information. All these parts are inter-connected. To model this database, the Merise method was utilised and 26 entities and 32 relationships were created. At the physical level, 93 tables were created, corresponding, for the majority, to entities and relationships of the data model. They were divided into reference tables (n = 65) and data tables (n = 28). Data processing was developed in Flora and included the control of the data, generic calculations of unknown data from given data, the automation of the estimation of the missing data or the chemical

  10. A kinetic and thermochemical database for organic sulfur and oxygen compounds.

    PubMed

    Class, Caleb A; Aguilera-Iparraguirre, Jorge; Green, William H

    2015-05-28

    Potential energy surfaces and reaction kinetics were calculated for 40 reactions involving sulfur and oxygen. This includes 11 H2O addition, 8 H2S addition, 11 hydrogen abstraction, 7 beta scission, and 3 elementary tautomerization reactions, which are potentially relevant in the combustion and desulfurization of sulfur compounds found in various fuel sources. Geometry optimizations and frequencies were calculated for reactants and transition states using B3LYP/CBSB7, and potential energies were calculated using CBS-QB3 and CCSD(T)-F12a/VTZ-F12. Rate coefficients were calculated using conventional transition state theory, with corrections for internal rotations and tunneling. Additionally, thermochemical parameters were calculated for each of the compounds involved in these reactions. With few exceptions, rate parameters calculated using the two potential energy methods agreed reasonably, with calculated activation energies differing by less than 5 kJ mol(-1). The computed rate coefficients and thermochemical parameters are expected to be useful for kinetic modeling. PMID:25941683

  11. Toward models of surgical procedures: analyzing a database of neurosurgical cases

    NASA Astrophysics Data System (ADS)

    Raimbault, Melanie; Morandi, Xavier; Jannin, Pierre

    2005-04-01

    Image-guided surgery systems can be improved by the knowledge of surgical expertise. The more the surgeon and the system know about the surgical procedure to perform beforehand, the easier it will be to plan and perform. The main objective of this paper is to introduce an approach for predicting surgical performance according to input variables related to the patient. This prediction is a first step towards the inclusion of surgical expertise in the image guided surgery systems. We previously proposed a generic model for describing surgical procedures in the specific context of multimodal neuronavigation. In this paper, we present the preliminary results of the analysis of a neurosurgical cases database built in accordance with the generic model and including 159 surgical cases concerning right-handed patients. We defined two queries on this surgical cases database to illustrate how it could be used to extract relevant and conclusive information about the surgical procedures: How does the anatomical localization of the target influence patient positioning? How does the anatomical localization of the target influence progress of the steps involved in the surgical procedure? The mid-term goal of our research is to semi automatically extract information, a priori models or scenarios of specific surgical procedures that can make easier the decision making process both for planning and surgery.

  12. Loops In Proteins (LIP)--a comprehensive loop database for homology modelling.

    PubMed

    Michalsky, E; Goede, A; Preissner, R

    2003-12-01

    One of the most important and challenging tasks in protein modelling is the prediction of loops, as can be seen in the large variety of existing approaches. Loops In Proteins (LIP) is a database that includes all protein segments of a length up to 15 residues contained in the Protein Data Bank (PDB). In this study, the applicability of LIP to loop prediction in the framework of homology modelling is investigated. Searching the database for loop candidates takes less than 1 s on a desktop PC, and ranking them takes a few minutes. This is an order of magnitude faster than most existing procedures. The measure of accuracy is the root mean square deviation (RMSD) with respect to the main-chain atoms after local superposition of target loop and predicted loop. Loops of up to nine residues length were modelled with a local RMSD <1 A and those of length up to 14 residues with an accuracy better than 2 A. The results were compared in detail with a thoroughly evaluated and tested ab initio method published recently and additionally with two further methods for a small loop test set. The LIP method produced very good predictions. In particular for longer loops it outperformed other methods.

  13. Ad HOC Model Generation Using Multiscale LIDAR Data from a Geospatial Database

    NASA Astrophysics Data System (ADS)

    Gordon, M.; Borgmann, B.; Gehrung, J.; Hebel, M.; Arens, M.

    2015-08-01

    Due to the spread of economically priced laser scanning technology nowadays, especially in the field of topographic surveying and mapping, ever-growing amounts of data need to be handled. Depending on the requirements of the specific application, airborne, mobile or terrestrial laser scanners are commonly used. Since visualizing this flood of data is not feasible with classical approaches like raw point cloud rendering, real time decision making requires sophisticated solutions. In addition, the efficient storage and recovery of 3D measurements is a challenging task. Therefore we propose an approach for the intelligent storage of 3D point clouds using a spatial database. For a given region of interest, the database is queried for the data available. All resulting point clouds are fused in a model generation process, utilizing the fact that low density airborne measurements could be used to supplement higher density mobile or terrestrial laser scans. The octree based modeling approach divides and subdivides the world into cells of varying size and fits one plane per cell, once a specified amount of points is present. The resulting model exceeds the completeness and precision of every single data source and enables for real time visualization. This is especially supported by data compression ratios of about 90%.

  14. Gyrokinetic Transport Database and Comparisons to the TGLF Theory-Based Transport Model

    NASA Astrophysics Data System (ADS)

    Kinsey, J. E.; Staebler, G. M.; Waltz, R. E.; Candy, J.

    2006-10-01

    A database with over 300 nonlinear gyrokinetic simulations has been created using the GYRO code [1,2]. Using a parameterized equilibrium model for shaped geometry, simulations show that the GYRO normalized ITG/TEM diffusivities exhibit an inverse linear dependence on elongation at fixed midplane minor radius. Kinetic electron simulations show the ExB shear quench rule is robust in shifted circle geometry. With real geometry, the quench point varies systematically with elongation and aspect ratio. Using the results, a new version of the quench rule is found that captures the variation of the quench point with these two geometric quantities. Finally, we compare the results from the TGLF driftwave model [3] with the GYRO simulations. Using the TGLF eigenmodes, we compute quasilinear fluxes using a turbulence saturation model and assess the quality of the fit to the GYRO transport database. 4pt[1] J. Candy, R.E. Waltz, Phys. Rev. Lett. 91, 45001 (2003). [2] http://fusion.gat.com/comp/parallel [3] G.M. Staebler, J.E. Kinsey, R.E. Waltz, Phys. Plasmas 12, 102508 (2005).

  15. A Bayesian Multivariate Receptor Model for Estimating Source Contributions to Particulate Matter Pollution using National Databases

    PubMed Central

    Hackstadt, Amber J.; Peng, Roger D.

    2014-01-01

    Summary Time series studies have suggested that air pollution can negatively impact health. These studies have typically focused on the total mass of fine particulate matter air pollution or the individual chemical constituents that contribute to it, and not source-specific contributions to air pollution. Source-specific contribution estimates are useful from a regulatory standpoint by allowing regulators to focus limited resources on reducing emissions from sources that are major contributors to air pollution and are also desired when estimating source-specific health effects. However, researchers often lack direct observations of the emissions at the source level. We propose a Bayesian multivariate receptor model to infer information about source contributions from ambient air pollution measurements. The proposed model incorporates information from national databases containing data on both the composition of source emissions and the amount of emissions from known sources of air pollution. The proposed model is used to perform source apportionment analyses for two distinct locations in the United States (Boston, Massachusetts and Phoenix, Arizona). Our results mirror previous source apportionment analyses that did not utilize the information from national databases and provide additional information about uncertainty that is relevant to the estimation of health effects. PMID:25309119

  16. InterMOD: integrated data and tools for the unification of model organism research

    PubMed Central

    Sullivan, Julie; Karra, Kalpana; Moxon, Sierra A. T.; Vallejos, Andrew; Motenko, Howie; Wong, J. D.; Aleksic, Jelena; Balakrishnan, Rama; Binkley, Gail; Harris, Todd; Hitz, Benjamin; Jayaraman, Pushkala; Lyne, Rachel; Neuhauser, Steven; Pich, Christian; Smith, Richard N.; Trinh, Quang; Cherry, J. Michael; Richardson, Joel; Stein, Lincoln; Twigger, Simon; Westerfield, Monte; Worthey, Elizabeth; Micklem, Gos

    2013-01-01

    Model organisms are widely used for understanding basic biology, and have significantly contributed to the study of human disease. In recent years, genomic analysis has provided extensive evidence of widespread conservation of gene sequence and function amongst eukaryotes, allowing insights from model organisms to help decipher gene function in a wider range of species. The InterMOD consortium is developing an infrastructure based around the InterMine data warehouse system to integrate genomic and functional data from a number of key model organisms, leading the way to improved cross-species research. So far including budding yeast, nematode worm, fruit fly, zebrafish, rat and mouse, the project has set up data warehouses, synchronized data models, and created analysis tools and links between data from different species. The project unites a number of major model organism databases, improving both the consistency and accessibility of comparative research, to the benefit of the wider scientific community. PMID:23652793

  17. InterMOD: integrated data and tools for the unification of model organism research.

    PubMed

    Sullivan, Julie; Karra, Kalpana; Moxon, Sierra A T; Vallejos, Andrew; Motenko, Howie; Wong, J D; Aleksic, Jelena; Balakrishnan, Rama; Binkley, Gail; Harris, Todd; Hitz, Benjamin; Jayaraman, Pushkala; Lyne, Rachel; Neuhauser, Steven; Pich, Christian; Smith, Richard N; Trinh, Quang; Cherry, J Michael; Richardson, Joel; Stein, Lincoln; Twigger, Simon; Westerfield, Monte; Worthey, Elizabeth; Micklem, Gos

    2013-01-01

    Model organisms are widely used for understanding basic biology, and have significantly contributed to the study of human disease. In recent years, genomic analysis has provided extensive evidence of widespread conservation of gene sequence and function amongst eukaryotes, allowing insights from model organisms to help decipher gene function in a wider range of species. The InterMOD consortium is developing an infrastructure based around the InterMine data warehouse system to integrate genomic and functional data from a number of key model organisms, leading the way to improved cross-species research. So far including budding yeast, nematode worm, fruit fly, zebrafish, rat and mouse, the project has set up data warehouses, synchronized data models, and created analysis tools and links between data from different species. The project unites a number of major model organism databases, improving both the consistency and accessibility of comparative research, to the benefit of the wider scientific community. PMID:23652793

  18. JAKs and STATs in invertebrate model organisms.

    PubMed

    Dearolf, C R

    1999-09-01

    Invertebrate organisms provide systems to elucidate the developmental roles of Janus kinase (JAK)/signal transducers and activators of transcription (STAT) signaling pathways, thereby complementing research conducted with mammalian cells and animals. Components of the JAK/STAT protein pathway have been identified and characterized in the fruit fly Drosophila melanogaster and the cellular slime mold Dictyostelium discoideum. This review summarizes the molecular and genetic data obtained from these model organisms. In particular, a Drosophila JAK/STAT pathway regulates normal segmentation, cell proliferation, and differentiation, and hyperactivation of the pathway leads to tumor formation and leukemia-like defects. A Dictyostelium STAT regulates the development of stalk cells during the multicellular part of the life cycle. Future research utilizing these organisms should continue to provide insights into the roles and regulation of these proteins and their signaling pathways. PMID:10526575

  19. Reflective Database Access Control

    ERIC Educational Resources Information Center

    Olson, Lars E.

    2009-01-01

    "Reflective Database Access Control" (RDBAC) is a model in which a database privilege is expressed as a database query itself, rather than as a static privilege contained in an access control list. RDBAC aids the management of database access controls by improving the expressiveness of policies. However, such policies introduce new interactions…

  20. Evaluation of a vortex-based subgrid stress model using DNS databases

    NASA Technical Reports Server (NTRS)

    Misra, Ashish; Lund, Thomas S.

    1996-01-01

    The performance of a SubGrid Stress (SGS) model for Large-Eddy Simulation (LES) developed by Misra k Pullin (1996) is studied for forced and decaying isotropic turbulence on a 32(exp 3) grid. The physical viability of the model assumptions are tested using DNS databases. The results from LES of forced turbulence at Taylor Reynolds number R(sub (lambda)) approximately equals 90 are compared with filtered DNS fields. Probability density functions (pdfs) of the subgrid energy transfer, total dissipation, and the stretch of the subgrid vorticity by the resolved velocity-gradient tensor show reasonable agreement with the DNS data. The model is also tested in LES of decaying isotropic turbulence where it correctly predicts the decay rate and energy spectra measured by Comte-Bellot & Corrsin (1971).

  1. Biomine: predicting links between biological entities using network models of heterogeneous databases

    PubMed Central

    2012-01-01

    Background Biological databases contain large amounts of data concerning the functions and associations of genes and proteins. Integration of data from several such databases into a single repository can aid the discovery of previously unknown connections spanning multiple types of relationships and databases. Results Biomine is a system that integrates cross-references from several biological databases into a graph model with multiple types of edges, such as protein interactions, gene-disease associations and gene ontology annotations. Edges are weighted based on their type, reliability, and informativeness. We present Biomine and evaluate its performance in link prediction, where the goal is to predict pairs of nodes that will be connected in the future, based on current data. In particular, we formulate protein interaction prediction and disease gene prioritization tasks as instances of link prediction. The predictions are based on a proximity measure computed on the integrated graph. We consider and experiment with several such measures, and perform a parameter optimization procedure where different edge types are weighted to optimize link prediction accuracy. We also propose a novel method for disease-gene prioritization, defined as finding a subset of candidate genes that cluster together in the graph. We experimentally evaluate Biomine by predicting future annotations in the source databases and prioritizing lists of putative disease genes. Conclusions The experimental results show that Biomine has strong potential for predicting links when a set of selected candidate links is available. The predictions obtained using the entire Biomine dataset are shown to clearly outperform ones obtained using any single source of data alone, when different types of links are suitably weighted. In the gene prioritization task, an established reference set of disease-associated genes is useful, but the results show that under favorable conditions, Biomine can also perform

  2. ADANS database specification

    SciTech Connect

    1997-01-16

    The purpose of the Air Mobility Command (AMC) Deployment Analysis System (ADANS) Database Specification (DS) is to describe the database organization and storage allocation and to provide the detailed data model of the physical design and information necessary for the construction of the parts of the database (e.g., tables, indexes, rules, defaults). The DS includes entity relationship diagrams, table and field definitions, reports on other database objects, and a description of the ADANS data dictionary. ADANS is the automated system used by Headquarters AMC and the Tanker Airlift Control Center (TACC) for airlift planning and scheduling of peacetime and contingency operations as well as for deliberate planning. ADANS also supports planning and scheduling of Air Refueling Events by the TACC and the unit-level tanker schedulers. ADANS receives input in the form of movement requirements and air refueling requests. It provides a suite of tools for planners to manipulate these requirements/requests against mobility assets and to develop, analyze, and distribute schedules. Analysis tools are provided for assessing the products of the scheduling subsystems, and editing capabilities support the refinement of schedules. A reporting capability provides formatted screen, print, and/or file outputs of various standard reports. An interface subsystem handles message traffic to and from external systems. The database is an integral part of the functionality summarized above.

  3. Modeling Secondary Organic Aerosol Formation From Emissions of Combustion Sources

    NASA Astrophysics Data System (ADS)

    Jathar, Shantanu Hemant

    Atmospheric aerosols exert a large influence on the Earth's climate and cause adverse public health effects, reduced visibility and material degradation. Secondary organic aerosol (SOA), defined as the aerosol mass arising from the oxidation products of gas-phase organic species, accounts for a significant fraction of the submicron atmospheric aerosol mass. Yet, there are large uncertainties surrounding the sources, atmospheric evolution and properties of SOA. This thesis combines laboratory experiments, extensive data analysis and global modeling to investigate the contribution of semi-volatile and intermediate volatility organic compounds (SVOC and IVOC) from combustion sources to SOA formation. The goals are to quantify the contribution of these emissions to ambient PM and to evaluate and improve models to simulate its formation. To create a database for model development and evaluation, a series of smog chamber experiments were conducted on evaporated fuel, which served as surrogates for real-world combustion emissions. Diesel formed the most SOA followed by conventional jet fuel / jet fuel derived from natural gas, gasoline and jet fuel derived from coal. The variability in SOA formation from actual combustion emissions can be partially explained by the composition of the fuel. Several models were developed and tested along with existing models using SOA data from smog chamber experiments conducted using evaporated fuel (this work, gasoline, fischertropschs, jet fuel, diesels) and published data on dilute combustion emissions (aircraft, on- and off-road gasoline, on- and off-road diesel, wood burning, biomass burning). For all of the SOA data, existing models under-predicted SOA formation if SVOC/IVOC were not included. For the evaporated fuel experiments, when SVOC/IVOC were included predictions using the existing SOA model were brought to within a factor of two of measurements with minor adjustments to model parameterizations. Further, a volatility

  4. Prediction model of potential hepatocarcinogenicity of rat hepatocarcinogens using a large-scale toxicogenomics database

    SciTech Connect

    Uehara, Takeki; Minowa, Yohsuke; Morikawa, Yuji; Kondo, Chiaki; Maruyama, Toshiyuki; Kato, Ikuo; Nakatsu, Noriyuki; Igarashi, Yoshinobu; Ono, Atsushi; Hayashi, Hitomi; Mitsumori, Kunitoshi; Yamada, Hiroshi; Ohno, Yasuo; Urushidani, Tetsuro

    2011-09-15

    The present study was performed to develop a robust gene-based prediction model for early assessment of potential hepatocarcinogenicity of chemicals in rats by using our toxicogenomics database, TG-GATEs (Genomics-Assisted Toxicity Evaluation System developed by the Toxicogenomics Project in Japan). The positive training set consisted of high- or middle-dose groups that received 6 different non-genotoxic hepatocarcinogens during a 28-day period. The negative training set consisted of high- or middle-dose groups of 54 non-carcinogens. Support vector machine combined with wrapper-type gene selection algorithms was used for modeling. Consequently, our best classifier yielded prediction accuracies for hepatocarcinogenicity of 99% sensitivity and 97% specificity in the training data set, and false positive prediction was almost completely eliminated. Pathway analysis of feature genes revealed that the mitogen-activated protein kinase p38- and phosphatidylinositol-3-kinase-centered interactome and the v-myc myelocytomatosis viral oncogene homolog-centered interactome were the 2 most significant networks. The usefulness and robustness of our predictor were further confirmed in an independent validation data set obtained from the public database. Interestingly, similar positive predictions were obtained in several genotoxic hepatocarcinogens as well as non-genotoxic hepatocarcinogens. These results indicate that the expression profiles of our newly selected candidate biomarker genes might be common characteristics in the early stage of carcinogenesis for both genotoxic and non-genotoxic carcinogens in the rat liver. Our toxicogenomic model might be useful for the prospective screening of hepatocarcinogenicity of compounds and prioritization of compounds for carcinogenicity testing. - Highlights: >We developed a toxicogenomic model to predict hepatocarcinogenicity of chemicals. >The optimized model consisting of 9 probes had 99% sensitivity and 97% specificity. >This model

  5. Charophytes: Evolutionary Giants and Emerging Model Organisms

    PubMed Central

    Domozych, David S.; Popper, Zoë A.; Sørensen, Iben

    2016-01-01

    Charophytes are the group of green algae whose ancestral lineage gave rise to land plants in what resulted in a profoundly transformative event in the natural history of the planet. Extant charophytes exhibit many features that are similar to those found in land plants and their relatively simple phenotypes make them efficacious organisms for the study of many fundamental biological phenomena. Several taxa including Micrasterias, Penium, Chara, and Coleochaete are valuable model organisms for the study of cell biology, development, physiology and ecology of plants. New and rapidly expanding molecular studies are increasing the use of charophytes that in turn, will dramatically enhance our understanding of the evolution of plants and the adaptations that allowed for survival on land. The Frontiers in Plant Science series on “Charophytes” provides an assortment of new research reports and reviews on charophytes and their emerging significance as model plants. PMID:27777578

  6. Subject and authorship of records related to the Organization for Tropical Studies (OTS) in BINABITROP, a comprehensive database about Costa Rican biology.

    PubMed

    Monge-Nájera, Julián; Nielsen-Muñoz, Vanessa; Azofeifa-Mora, Ana Beatriz

    2013-06-01

    BINABITROP is a bibliographical database of more than 38000 records about the ecosystems and organisms of Costa Rica. In contrast with commercial databases, such as Web of Knowledge and Scopus, which exclude most of the scientific journals published in tropical countries, BINABITROP is a comprehensive record of knowledge on the tropical ecosystems and organisms of Costa Rica. We analyzed its contents in three sites (La Selva, Palo Verde and Las Cruces) and recorded scientific field, taxonomic group and authorship. We found that most records dealt with ecology and systematics, and that most authors published only one article in the study period (1963-2011). Most research was published in four journals: Biotropica, Revista de Biología Tropical/ International Journal of Tropical Biology and Conservation, Zootaxa and Brenesia. This may be the first study of a such a comprehensive database for any case of tropical biology literature.

  7. Multiscale generalized Born modeling of ligand binding energies for virtual database screening

    PubMed Central

    Liu, Hao-Yang; Grinter, Sam Z.; Zou, Xiaoqin

    2009-01-01

    Generalized Born (GB) models are widely used to study the electrostatic energetics of solute molecules including proteins. Previous work demonstrates that GB models may produce satisfactory solvation energies if accurate effective Born radii are computed for all atoms. Our previous study showed that a GB model which reproduces the solvation energy may not necessarily be suitable for ligand binding calculations. In this work, we studied binding energetics using the exact GB model, in which Born radii are computed from the Poisson-Boltzmann (PB) equation. Our results showed that accurate Born radii lead to very good agreement between GB and PB in electrostatic calculations for ligand binding. However, recently developed GB models with high Born radii accuracy, when used in large database screening, may suffer from time constraints which make accurate, large-scale Born radii calculations impractical. We therefore present a multiscale GB approach in which atoms are divided into two groups. For atoms in the first group, those few atoms which are most likely to be critical to binding electrostatics, the Born radii are computed accurately at the sacrifice of speed. We propose two alternative approaches for atoms in the second group. The Born radii of these atoms may simply be computed by a fast GB method. Alternatively, the Born radii of these atoms may be computed accurately in the free state, then a variational form of a fast GB method may be used to compute the change in Born radii experienced by these atoms during binding. This strategy provides an accuracy advantage while still being fast enough for use in the virtual screening of large databases. PMID:19678651

  8. A Conceptual Model and Database to Integrate Data and Project Management

    NASA Astrophysics Data System (ADS)

    Guarinello, M. L.; Edsall, R.; Helbling, J.; Evaldt, E.; Glenn, N. F.; Delparte, D.; Sheneman, L.; Schumaker, R.

    2015-12-01

    Data management is critically foundational to doing effective science in our data-intensive research era and done well can enhance collaboration, increase the value of research data, and support requirements by funding agencies to make scientific data and other research products available through publically accessible online repositories. However, there are few examples (but see the Long-term Ecological Research Network Data Portal) of these data being provided in such a manner that allows exploration within the context of the research process - what specific research questions do these data seek to answer? what data were used to answer these questions? what data would have been helpful to answer these questions but were not available? We propose an agile conceptual model and database design, as well as example results, that integrate data management with project management not only to maximize the value of research data products but to enhance collaboration during the project and the process of project management itself. In our project, which we call 'Data Map,' we used agile principles by adopting a user-focused approach and by designing our database to be simple, responsive, and expandable. We initially designed Data Map for the Idaho EPSCoR project "Managing Idaho's Landscapes for Ecosystem Services (MILES)" (see https://www.idahoecosystems.org//) and will present example results for this work. We consulted with our primary users- project managers, data managers, and researchers to design the Data Map. Results will be useful to project managers and to funding agencies reviewing progress because they will readily provide answers to the questions "For which research projects/questions are data available and/or being generated by MILES researchers?" and "Which research projects/questions are associated with each of the 3 primary questions from the MILES proposal?" To be responsive to the needs of the project, we chose to streamline our design for the prototype

  9. Assessment of methods for creating a national building statistics database for atmospheric dispersion modeling

    SciTech Connect

    Velugubantla, S. P.; Burian, S. J.; Brown, M. J.; McKinnon, A. T.; McPherson, T. N.; Han, W. S.

    2004-01-01

    Mesoscale meteorological codes and transport and dispersion models are increasingly being applied in urban areas. Representing urban terrain characteristics in these models is critical for accurate predictions of air flow, heating and cooling, and airborne contaminant concentrations in cities. A key component of urban terrain characterization is the description of building morphology (e.g., height, plan area, frontal area) and derived properties (e.g., roughness length). Methods to determine building morphological statistics range from manual field surveys to automated processing of digital building databases. In order to improve the quality and consistency of mesoscale meteorological and atmospheric dispersion modeling, a national dataset of building morphological statistics is needed. Currently, due to the expense and logistics of conducting detailed field surveys, building statistics have been derived for only small sections of a few cities. In most other cities, modeling projects rely on building statistics estimated using intuition and best guesses. There has been increasing emphasis in recent years to derive building statistics using digital building data or other data sources as a proxy for those data. Although there is a current expansion in public and private sector development of digital building data, at present there is insufficient data to derive a national building statistics database using automated analysis tools. Too many cities lack digital data on building footprints and heights and many of the cities having such data do so for only small areas. Due to the lack of sufficient digital building data, other datasets are used to estimate building statistics. Land use often serves as means to provide building statistics for a model domain, but the strength and consistency of the relationship between land use and building morphology is largely uncertain. In this paper, we investigate whether building statistics can be correlated to the underlying land

  10. High resolution topography and land cover databases for wind resource assessment using mesoscale models

    NASA Astrophysics Data System (ADS)

    Barranger, Nicolas; Stathopoulos, Christos; Kallos, Georges

    2013-04-01

    In wind resource assessment, mesoscale models can provide wind flow characteristics without the use of mast measurements. In complex terrain, local orography and land cover data assimilation are essential parameters to accurately simulate the wind flow pattern within the atmospheric boundary layer. State-of-the-art Mesoscale Models such as RAMS usually provides orography and landuse data with of resolution of 30s (about 1km). This resolution is necessary for solving mesocale phenomena accurately but not sufficient when the aim is to quantitatively estimate the wind flow characteristics passing over sharp hills or ridges. Furthermore, the abrupt change in land cover characterization is nor always taken into account in the model with a low resolution land use database. When land cover characteristics changes dramatically, parameters such as roughness, albedo or soil moisture that can highly influence the Atmospheric Boundary Layer meteorological characteristics. Therefore they require to be accurately assimilated into the model. Since few years, high resolution databases derived from satellite imagery (Modis, SRTM, LandSat, SPOT ) are available online. Being converted to RAMS requirements inputs, an evaluation of the model requires to be achieved. For this purpose, three new high resolution land cover and two topographical databases are implemented and tested in RAMS. The analysis of terrain variability is performed using basis functions of space frequency and amplitude. Practically, one and two dimension Fast Fourier Transform is applied to terrain height to reveal the main characteristics of local orography according to the obtained wave spectrum. By this way, a comparison between different topographic data sets is performed, based on the terrain power spectrum entailed in the terrain height input. Furthermore, this analysis is a powerful tool in the determination of the proper horizontal grid resolution required to resolve most of the energy containing spectrum

  11. Modelling motions within the organ of Corti

    NASA Astrophysics Data System (ADS)

    Ni, Guangjian; Baumgart, Johannes; Elliott, Stephen

    2015-12-01

    Most cochlear models used to describe the basilar membrane vibration along the cochlea are concerned with macromechanics, and often assume that the organ of Corti moves as a single unit, ignoring the individual motion of different components. New experimental technologies provide the opportunity to measure the dynamic behaviour of different components within the organ of Corti, but only for certain types of excitation. It is thus still difficult to directly measure every aspect of cochlear dynamics, particularly for acoustic excitation of the fully active cochlea. The present work studies the dynamic response of a model of the cross-section of the cochlea, at the microscopic level, using the finite element method. The elastic components are modelled with plate elements and the perilymph and endolymph are modelled with inviscid fluid elements. The individual motion of each component within the organ of Corti is calculated with dynamic pressure loading on the basilar membrane and the motions of the experimentally accessible parts are compared with measurements. The reticular lamina moves as a stiff plate, without much bending, and is pivoting around a point close to the region of the inner hair cells, as observed experimentally. The basilar membrane shows a slightly asymmetric mode shape, with maximum displacement occurring between the second-row and the third-row of the outer hair cells. The dynamics responses is also calculated, and compared with experiments, when driven by the outer hair cells. The receptance of the basilar membrane motion and of the deflection of the hair bundles of the outer hair cells is thus obtained, when driven either acoustically or electrically. In this way, the fully active linear response of the basilar membrane to acoustic excitation can be predicted by using a linear superposition of the calculated receptances and a defined gain function for the outer hair cell feedback.

  12. The NCBI Taxonomy database.

    PubMed

    Federhen, Scott

    2012-01-01

    The NCBI Taxonomy database (http://www.ncbi.nlm.nih.gov/taxonomy) is the standard nomenclature and classification repository for the International Nucleotide Sequence Database Collaboration (INSDC), comprising the GenBank, ENA (EMBL) and DDBJ databases. It includes organism names and taxonomic lineages for each of the sequences represented in the INSDC's nucleotide and protein sequence databases. The taxonomy database is manually curated by a small group of scientists at the NCBI who use the current taxonomic literature to maintain a phylogenetic taxonomy for the source organisms represented in the sequence databases. The taxonomy database is a central organizing hub for many of the resources at the NCBI, and provides a means for clustering elements within other domains of NCBI web site, for internal linking between domains of the Entrez system and for linking out to taxon-specific external resources on the web. Our primary purpose is to index the domain of sequences as conveniently as possible for our user community.

  13. Collection Fusion Using Bayesian Estimation of a Linear Regression Model in Image Databases on the Web.

    ERIC Educational Resources Information Center

    Kim, Deok-Hwan; Chung, Chin-Wan

    2003-01-01

    Discusses the collection fusion problem of image databases, concerned with retrieving relevant images by content based retrieval from image databases distributed on the Web. Focuses on a metaserver which selects image databases supporting similarity measures and proposes a new algorithm which exploits a probabilistic technique using Bayesian…

  14. MODEL-BASED HYDROACOUSTIC BLOCKAGE ASSESSMENT AND DEVELOPMENT OF AN EXPLOSIVE SOURCE DATABASE

    SciTech Connect

    Matzel, E; Ramirez, A; Harben, P

    2005-07-11

    We are continuing the development of the Hydroacoustic Blockage Assessment Tool (HABAT) which is designed for use by analysts to predict which hydroacoustic monitoring stations can be used in discrimination analysis for any particular event. The research involves two approaches (1) model-based assessment of blockage, and (2) ground-truth data-based assessment of blockage. The tool presents the analyst with a map of the world, and plots raypath blockages from stations to sources. The analyst inputs source locations and blockage criteria, and the tool returns a list of blockage status from all source locations to all hydroacoustic stations. We are currently using the tool in an assessment of blockage criteria for simple direct-path arrivals. Hydroacoustic data, predominantly from earthquake sources, are read in and assessed for blockage at all available stations. Several measures are taken. First, can the event be observed at a station above background noise? Second, can we establish backazimuth from the station to the source. Third, how large is the decibel drop at one station relative to other stations. These observational results are then compared with model estimates to identify the best set of blockage criteria and used to create a set of blockage maps for each station. The model-based estimates are currently limited by the coarse bathymetry of existing databases and by the limitations inherent in the raytrace method. In collaboration with BBN Inc., the Hydroacoustic Coverage Assessment Model (HydroCAM) that generates the blockage files that serve as input to HABAT, is being extended to include high-resolution bathymetry databases in key areas that increase model-based blockage assessment reliability. An important aspect of this capability is to eventually include reflected T-phases where they reliably occur and to identify the associated reflectors. To assess how well any given hydroacoustic discriminant works in separating earthquake and in-water explosion

  15. A neotropical Miocene pollen database employing image-based search and semantic modeling1

    PubMed Central

    Han, Jing Ginger; Cao, Hongfei; Barb, Adrian; Punyasena, Surangi W.; Jaramillo, Carlos; Shyu, Chi-Ren

    2014-01-01

    • Premise of the study: Digital microscopic pollen images are being generated with increasing speed and volume, producing opportunities to develop new computational methods that increase the consistency and efficiency of pollen analysis and provide the palynological community a computational framework for information sharing and knowledge transfer. • Methods: Mathematical methods were used to assign trait semantics (abstract morphological representations) of the images of neotropical Miocene pollen and spores. Advanced database-indexing structures were built to compare and retrieve similar images based on their visual content. A Web-based system was developed to provide novel tools for automatic trait semantic annotation and image retrieval by trait semantics and visual content. • Results: Mathematical models that map visual features to trait semantics can be used to annotate images with morphology semantics and to search image databases with improved reliability and productivity. Images can also be searched by visual content, providing users with customized emphases on traits such as color, shape, and texture. • Discussion: Content- and semantic-based image searches provide a powerful computational platform for pollen and spore identification. The infrastructure outlined provides a framework for building a community-wide palynological resource, streamlining the process of manual identification, analysis, and species discovery. PMID:25202648

  16. The Traumatic Brain Injury Model Systems: a longitudinal database, research, collaboration and knowledge translation.

    PubMed

    Hammond, F M; Malec, J F

    2010-12-01

    In 1988, the National Institute on Disability and Rehabilitation Research (NIDRR) launched the Traumatic Brain Injury Model Systems (TBIMS) program, creating the longest and largest longitudinal database on individuals with moderate-to-severe traumatic brain injury (TBI) available today. In addition to sustaining the longitudinal database, centers that successfully compete to be part of the TBIMS centers are also expected to complete local and collaborative research projects to further scientific knowledge about TBI. The research has focused on areas of the NIDRR Long Range Plan which emphasizes employment, health and function, technology for access and function, independent living and community integration, and other associated disability research areas. Centers compete for funded participation in the TBIMS on a 5-year cycle. Dissemination of scientific knowledge gained through the TBIMS is the responsibility of both individual centers and the TBIMS as a whole. This is accomplished through multiple venues that target a broad audience of those who need to receive the information and learn how to best apply it to practice. The sites produce many useful websites, manuals, publications and other materials to accomplish this translation of knowledge to practice.

  17. Meta-Analysis in Human Neuroimaging: Computational Modeling of Large-Scale Databases

    PubMed Central

    Fox, Peter T.; Lancaster, Jack L.; Laird, Angela R.; Eickhoff, Simon B.

    2016-01-01

    Spatial normalization—applying standardized coordinates as anatomical addresses within a reference space—was introduced to human neuroimaging research nearly 30 years ago. Over these three decades, an impressive series of methodological advances have adopted, extended, and popularized this standard. Collectively, this work has generated a methodologically coherent literature of unprecedented rigor, size, and scope. Large-scale online databases have compiled these observations and their associated meta-data, stimulating the development of meta-analytic methods to exploit this expanding corpus. Coordinate-based meta-analytic methods have emerged and evolved in rigor and utility. Early methods computed cross-study consensus, in a manner roughly comparable to traditional (nonimaging) meta-analysis. Recent advances now compute coactivation-based connectivity, connectivity-based functional parcellation, and complex network models powered from data sets representing tens of thousands of subjects. Meta-analyses of human neuroimaging data in large-scale databases now stand at the forefront of computational neurobiology. PMID:25032500

  18. A One-Degree Seismic Tomographic Model Based on a Sensitivity Kernel Database

    NASA Astrophysics Data System (ADS)

    Sales de Andrade, E.; Liu, Q.; Manners, U.; Lee-Varisco, E.; Ma, Z.; Masters, G.

    2013-12-01

    Seismic tomography is instrumental in mapping 3D velocity structures of the Earth's interior based on travel-time measurements and waveform differences. Although both ray theory and other asymptotic methods have been successfully employed in global tomography, they are less accurate for long-period waves or steep velocity gradients. They also lack the ability to predict 'non-geometrical' effects such as those for the core diffracted phases (Pdiff, Sdiff) which are crucial for mapping heterogeneities in the lowermost mantle (D'' layer). On the other hand, sensitivity kernels can be accurately calculated with no approximations by the interaction of forward and adjoint wavefields, both numerically simulated by spectral element methods. We have previously shown that by taking advantage of the symmetry of 1D reference models, we can efficiently and speedily construct sensitivity kernels of both P and S wavespeeds based on the simulation and storage of forward and adjoint strain fields for select source and receiver geometries. This technique has been used to create a database of strain fields as well as sensitivity kernels for phases typically used in global inversions. We also performed picks for 27,000 Sdiff, 35,000 Pdiff, 400,000 S, and 600,000 P phases and 33,000 SS-S, 33,000 PP-P, and 41,000 ScS-S differential phases, which provide much improved coverage of the globe. Using these travel-times and our sensitivity kernel database in a parallel LSQR inversion, we generate an updated tomographic model with 1° resolution. Using this improved coverage, we investigate differences between global models inverted based on ray theory and finite-frequency kernels.

  19. Automated clustering of ensembles of alternative models in protein structure databases.

    PubMed

    Domingues, Francisco S; Rahnenführer, Jörg; Lengauer, Thomas

    2004-06-01

    Experimentally determined protein structures have been classified in different public databases according to their structural and evolutionary relationships. Frequently, alternative structural models, determined using X-ray crystallography or NMR spectroscopy, are available for a protein. These models can present significant structural dissimilarity. Currently there is no classification available for these alternative structures. In order to classify them, we developed STRuster, an automated method for clustering ensembles of structural models according to their backbone structure. The method is based on the calculation of carbon alpha (Calpha) distance matrices. Two filters are applied in the calculation of the dissimilarity measure in order to identify both large and small (but significant) backbone conformational changes. The resulting dissimilarity value is used for hierarchical clustering and partitioning around medoids (PAM). Hierarchical clustering reflects the hierarchy of similarities between all pairs of models, while PAM groups the models into the 'optimal' number of clusters. The method has been applied to cluster the structures in each SCOP species level and can be easily applied to any other sets of conformers. The results are available at: http://bioinf.mpi-sb.mpg.de/projects/struster/. PMID:15319469

  20. Studying Oogenesis in a Non-model Organism Using Transcriptomics: Assembling, Annotating, and Analyzing Your Data.

    PubMed

    Carter, Jean-Michel; Gibbs, Melanie; Breuker, Casper J

    2016-01-01

    This chapter provides a guide to processing and analyzing RNA-Seq data in a non-model organism. This approach was implemented for studying oogenesis in the Speckled Wood Butterfly Pararge aegeria. We focus in particular on how to perform a more informative primary annotation of your non-model organism by implementing our multi-BLAST annotation strategy. We also provide a general guide to other essential steps in the next-generation sequencing analysis workflow. Before undertaking these methods, we recommend you familiarize yourself with command line usage and fundamental concepts of database handling. Most of the operations in the primary annotation pipeline can be performed in Galaxy (or equivalent standalone versions of the tools) and through the use of common database operations (e.g. to remove duplicates) but other equivalent programs and/or custom scripts can be implemented for further automation. PMID:27557578

  1. Lipidomic Profiling of Model Organisms and the World’s Major Pathogens

    PubMed Central

    Layre, Emilie; Moody, David Branch

    2014-01-01

    Lipidomics is a subspecialty of metabolomics that focuses on water insoluble metabolites that form membrane barriers. Most lipidomic databases catalog lipids from common model organisms, like humans or Escherichia coli. However, model organisms’ lipid profiles show surprisingly little overlap with that of specialized pathogens, creating the need for organism-specific lipidomic databases. Here we review rapid progress in lipidomic platform development with regard to chromatography, detection and bioinformatics. We emphasize new methods of comparative lipidomics, which use aligned datasets to identify lipids changed after introducing a biological variable. These new methods provide an unprecedented ability to broadly and quantitatively describe lipidic change during biological processes and identify changed lipids with low error rates. PMID:22971440

  2. Making designer mutants in model organisms.

    PubMed

    Peng, Ying; Clark, Karl J; Campbell, Jarryd M; Panetta, Magdalena R; Guo, Yi; Ekker, Stephen C

    2014-11-01

    Recent advances in the targeted modification of complex eukaryotic genomes have unlocked a new era of genome engineering. From the pioneering work using zinc-finger nucleases (ZFNs), to the advent of the versatile and specific TALEN systems, and most recently the highly accessible CRISPR/Cas9 systems, we now possess an unprecedented ability to analyze developmental processes using sophisticated designer genetic tools. In this Review, we summarize the common approaches and applications of these still-evolving tools as they are being used in the most popular model developmental systems. Excitingly, these robust and simple genomic engineering tools also promise to revolutionize developmental studies using less well established experimental organisms.

  3. TI: The West Coast and Alaska Tsunami Warning Center Forecast Model Project Applied to an Operational Tsunami Threat-database

    NASA Astrophysics Data System (ADS)

    Knight, W.; Huang, P.; Whitmore, P.; Sterling, K.

    2008-12-01

    Continuous improvement in the NOAA/West Coast & Alaska Tsunami Warning Center (WCATWC) forecast model has allowed the consideration of new uses for this model. These improvements include a finer propagation mesh, more model sources and magnitudes, runup boundary conditions, and continuous, unbroken fine coastal meshes. The focus of this report is on a new operational use of the model at the WCATWC - creation of a threat database of tsunami impacts on US and Canadian coastlines. Since all forecast model data is pre-computed, this concept should be easily realized. One recent case which showed the utility of a model-based threat database was the 4-1-2007 Solomon Islands Tsunami event. Tsunami energy maps clearly showed the energy was directed southwest and was no danger to regions to the northeast. Another case was the use of modeled tsunamis and their synthetic mareograms in the design of Gulf and Atlantic coast tsunami warning criteria. Currently, the only quantitative model data to appear in tsunami messages are ETAs for the leading edge of the tsunami wave train (the expected impact level is described in text - based on forecast model data). Since runups can now be forecasted for any coastal point, they can be used to constrain initial warning/watch/advisory messages to only threatened regions and can be saved to a database for later inclusion (along with ETAs) in tsunami bulletins. Present practice is to include all areas within a certain travel time or distance from epicenter in the initial warning bulletin, regardless of the threat. Since watch-warning- advisory breakpoints are based in the later bulletins on forecasted wave heights, the database can also be used to refine the extent of the warned zones. With full modeled mareograms similarly saved to a database, additional wave information like initial recession / elevation, or ETAs for first and highest waves can be added to tsunami bulletins. By comparison of scaled model prediction to historic tide gauge

  4. Integrated Standardized Database/Model Management System: Study management concepts and requirements

    SciTech Connect

    Baker, R.; Swerdlow, S.; Schultz, R.; Tolchin, R.

    1994-02-01

    Data-sharing among planners and planning software for utility companies is the motivation for creating the Integrated Standardized Database (ISD) and Model Management System (MMS). The purpose of this document is to define the requirements for the ISD/MMS study management component in a manner that will enhance the use of the ISD. After an analysis period which involved EPRI member utilities across the United States, the study concept was formulated. It is defined in terms of its entities, relationships and its support processes, specifically for implementation as the key component of the MMS. From the study concept definition, requirements are derived. There are unique requirements, such as the necessity to interface with DSManager, EGEAS, IRPManager, MIDAS and UPM and there are standard information systems requirements, such as create, modify, delete and browse data. An initial ordering of the requirements is established, with a section devoted to future enhancements.

  5. Spinal Cord Injury Model Systems: Review of Program and National Database From 1970 to 2015.

    PubMed

    Chen, Yuying; DeVivo, Michael J; Richards, J Scott; SanAgustin, Theresa B

    2016-10-01

    The Spinal Cord Injury Model Systems (SCIMS) centers have provided continuous, comprehensive multidisciplinary care for persons with spinal cord injury (SCI) in the United States since their inception in 1970. In addition, the research conducted and the analysis of data collected at these centers facilitate advances in the care and the overall quality of life for people with SCI. Over the past 45 years, the SCIMS program and National Spinal Cord Injury Database (NSCID) have undergone major revisions, which must be recognized in the planning, conduct, and interpretation of SCIMS research to prevent misinterpretation of findings. Therefore, we provide herein a brief review of the SCIMS program and the associated NSCID throughout its history, emphasizing changes and accomplishments within the past 15 years, to facilitate a better understanding and interpretation of the data presented in SCIMS research publications, including the articles published in this special issue of the Archives. PMID:27671806

  6. Engineering the object-relation database model in O-Raid

    NASA Technical Reports Server (NTRS)

    Dewan, Prasun; Vikram, Ashish; Bhargava, Bharat

    1989-01-01

    Raid is a distributed database system based on the relational model. O-raid is an extension of the Raid system and will support complex data objects. The design of O-Raid is evolutionary and retains all features of relational data base systems and those of a general purpose object-oriented programming language. O-Raid has several novel properties. Objects, classes, and inheritance are supported together with a predicate-base relational query language. O-Raid objects are compatible with C++ objects and may be read and manipulated by a C++ program without any 'impedance mismatch'. Relations and columns within relations may themselves be treated as objects with associated variables and methods. Relations may contain heterogeneous objects, that is, objects of more than one class in a certain column, which can individually evolve by being reclassified. Special facilities are provided to reduce the data search in a relation containing complex objects.

  7. Autophagy and cell death in model organisms.

    PubMed

    Kourtis, N; Tavernarakis, N

    2009-01-01

    Autophagy evolved in unicellular eukaryotes as a means for surviving nutrient stress. During the course of evolution, as multicellular organisms developed specialized cell types and complex intracellular signalling networks, autophagy has been summoned to serve additional cellular functions. Numerous recent studies indicate that apart from its pro-survival role under nutrient limitation, autophagy also participates in cell death. However, the precise role of this catabolic process in dying cells is not fully understood. Although in certain situations autophagy has a protective function, in other types of cell death it actually contributes to cellular destruction. Simple model organisms ranging from the unicellular Saccharomyces cerevisiae to the soil amoeba Dictyostelium discoideum and the metazoans Caenorhabditis elegans and Drosophila melanogaster provide clearly defined cell death paradigms that can be used to dissect the involvement of autophagy in cell death, at the molecular level. In this review, we survey current research in simple organisms, linking autophagy to cell death and discuss the complex interplay between autophagy, cell survival and cell death. PMID:19079286

  8. MEMOSys 2.0: an update of the bioinformatics database for genome-scale models and genomic data.

    PubMed

    Pabinger, Stephan; Snajder, Rene; Hardiman, Timo; Willi, Michaela; Dander, Andreas; Trajanoski, Zlatko

    2014-01-01

    The MEtabolic MOdel research and development System (MEMOSys) is a versatile database for the management, storage and development of genome-scale models (GEMs). Since its initial release, the database has undergone major improvements, and the new version introduces several new features. First, the novel concept of derived models allows users to create model hierarchies that automatically propagate modifications along their order. Second, all stored components can now be easily enhanced with additional annotations that can be directly extracted from a supplied Systems Biology Markup Language (SBML) file. Third, the web application has been substantially revised and now features new query mechanisms, an easy search system for reactions and new link-out services to publicly available databases. Fourth, the updated database now contains 20 publicly available models, which can be easily exported into standardized formats for further analysis. Fifth, MEMOSys 2.0 is now also available as a fully configured virtual image and can be found online at http://www.icbi.at/memosys and http://memoys.i-med.ac.at. Database URL: http://memosys.i-med.ac.at.

  9. Volcanogenic Massive Sulfide Deposits of the World - Database and Grade and Tonnage Models

    USGS Publications Warehouse

    Mosier, Dan L.; Berger, Vladimir I.; Singer, Donald A.

    2009-01-01

    Grade and tonnage models are useful in quantitative mineral-resource assessments. The models and database presented in this report are an update of earlier publications about volcanogenic massive sulfide (VMS) deposits. These VMS deposits include what were formerly classified as kuroko, Cyprus, and Besshi deposits. The update was necessary because of new information about some deposits, changes in information in some deposits, such as grades, tonnages, or ages, revised locations of some deposits, and reclassification of subtypes. In this report we have added new VMS deposits and removed a few incorrectly classified deposits. This global compilation of VMS deposits contains 1,090 deposits; however, it was not our intent to include every known deposit in the world. The data was recently used for mineral-deposit density models (Mosier and others, 2007; Singer, 2008). In this paper, 867 deposits were used to construct revised grade and tonnage models. Our new models are based on a reclassification of deposits based on host lithologies: Felsic, Bimodal-Mafic, and Mafic volcanogenic massive sulfide deposits. Mineral-deposit models are important in exploration planning and quantitative resource assessments for two reasons: (1) grades and tonnages among deposit types vary significantly, and (2) deposits of different types occur in distinct geologic settings that can be identified from geologic maps. Mineral-deposit models combine the diverse geoscience information on geology, mineral occurrences, geophysics, and geochemistry used in resource assessments and mineral exploration. Globally based deposit models allow recognition of important features and demonstrate how common different features are. Well-designed deposit models allow geologists to deduce possible mineral-deposit types in a given geologic environment and economists to determine the possible economic viability of these resources. Thus, mineral-deposit models play a central role in presenting geoscience

  10. The mouse genome database (MGD): new features facilitating a model system.

    PubMed

    Eppig, Janan T; Blake, Judith A; Bult, Carol J; Kadin, James A; Richardson, Joel E

    2007-01-01

    The mouse genome database (MGD, http://www.informatics.jax.org/), the international community database for mouse, provides access to extensive integrated data on the genetics, genomics and biology of the laboratory mouse. The mouse is an excellent and unique animal surrogate for studying normal development and disease processes in humans. Thus, MGD's primary goals are to facilitate the use of mouse models for studying human disease and enable the development of translational research hypotheses based on comparative genotype, phenotype and functional analyses. Core MGD data content includes gene characterization and functions, phenotype and disease model descriptions, DNA and protein sequence data, polymorphisms, gene mapping data and genome coordinates, and comparative gene data focused on mammals. Data are integrated from diverse sources, ranging from major resource centers to individual investigator laboratories and the scientific literature, using a combination of automated processes and expert human curation. MGD collaborates with the bioinformatics community on the development of data and semantic standards, and it incorporates key ontologies into the MGD annotation system, including the Gene Ontology (GO), the Mammalian Phenotype Ontology, and the Anatomical Dictionary for Mouse Development and the Adult Anatomy. MGD is the authoritative source for mouse nomenclature for genes, alleles, and mouse strains, and for GO annotations to mouse genes. MGD provides a unique platform for data mining and hypothesis generation where one can express complex queries simultaneously addressing phenotypic effects, biochemical function and process, sub-cellular location, expression, sequence, polymorphism and mapping data. Both web-based querying and computational access to data are provided. Recent improvements in MGD described here include the incorporation of single nucleotide polymorphism data and search tools, the addition of PIR gene superfamily classifications

  11. Drug-target interaction prediction: databases, web servers and computational models.

    PubMed

    Chen, Xing; Yan, Chenggang Clarence; Zhang, Xiaotian; Zhang, Xu; Dai, Feng; Yin, Jian; Zhang, Yongdong

    2016-07-01

    Identification of drug-target interactions is an important process in drug discovery. Although high-throughput screening and other biological assays are becoming available, experimental methods for drug-target interaction identification remain to be extremely costly, time-consuming and challenging even nowadays. Therefore, various computational models have been developed to predict potential drug-target associations on a large scale. In this review, databases and web servers involved in drug-target identification and drug discovery are summarized. In addition, we mainly introduced some state-of-the-art computational models for drug-target interactions prediction, including network-based method, machine learning-based method and so on. Specially, for the machine learning-based method, much attention was paid to supervised and semi-supervised models, which have essential difference in the adoption of negative samples. Although significant improvements for drug-target interaction prediction have been obtained by many effective computational models, both network-based and machine learning-based methods have their disadvantages, respectively. Furthermore, we discuss the future directions of the network-based drug discovery and network approach for personalized drug discovery based on personalized medicine, genome sequencing, tumor clone-based network and cancer hallmark-based network. Finally, we discussed the new evaluation validation framework and the formulation of drug-target interactions prediction problem by more realistic regression formulation based on quantitative bioactivity data.

  12. Combining a weed traits database with a population dynamics model predicts shifts in weed communities

    PubMed Central

    Storkey, J; Holst, N; Bøjer, O Q; Bigongiali, F; Bocci, G; Colbach, N; Dorner, Z; Riemens, M M; Sartorato, I; Sønderskov, M; Verschwele, A

    2015-01-01

    A functional approach to predicting shifts in weed floras in response to management or environmental change requires the combination of data on weed traits with analytical frameworks that capture the filtering effect of selection pressures on traits. A weed traits database (WTDB) was designed, populated and analysed, initially using data for 19 common European weeds, to begin to consolidate trait data in a single repository. The initial choice of traits was driven by the requirements of empirical models of weed population dynamics to identify correlations between traits and model parameters. These relationships were used to build a generic model, operating at the level of functional traits, to simulate the impact of increasing herbicide and fertiliser use on virtual weeds along gradients of seed weight and maximum height. The model generated ‘fitness contours’ (defined as population growth rates) within this trait space in different scenarios, onto which two sets of weed species, defined as common or declining in the UK, were mapped. The effect of increasing inputs on the weed flora was successfully simulated; 77% of common species were predicted to have stable or increasing populations under high fertiliser and herbicide use, in contrast with only 29% of the species that have declined. Future development of the WTDB will aim to increase the number of species covered, incorporate a wider range of traits and analyse intraspecific variability under contrasting management and environments. PMID:26190870

  13. Human intestinal transporter database: QSAR modeling and virtual profiling of drug uptake, efflux and interactions

    PubMed Central

    Sedykh, Alexander; Fourches, Denis; Duan, Jianmin; Hucke, Oliver; Garneau, Michel; Zhu, Hao; Bonneau, Pierre; Tropsha, Alexander

    2013-01-01

    Purpose Membrane transporters mediate many biological effects of chemicals and play a major role in pharmacokinetics and drug resistance. The selection of viable drug candidates among biologically active compounds requires the assessment of their transporter interaction profiles. Methods Using public sources, we have assembled and curated the largest, to our knowledge, human intestinal transporter database (>5,000 interaction entries for >3,700 molecules). This data was used to develop thoroughly validated classification Quantitative Structure-Activity Relationship (QSAR) models of transport and/or inhibition of several major transporters including MDR1, BCRP, MRP1-4, PEPT1, ASBT, OATP2B1, OCT1, and MCT1. Results & Conclusions QSAR models have been developed with advanced machine learning techniques such as Support Vector Machines, Random Forest, and k Nearest Neighbors using Dragon and MOE chemical descriptors. These models afforded high external prediction accuracies of 71–100% estimated by 5-fold external validation, and showed hit retrieval rates with up to 20-fold enrichment in the virtual screening of DrugBank compounds. The compendium of predictive QSAR models developed in this study can be used for virtual profiling of drug candidates and/or environmental agents with the optimal transporter profiles. PMID:23269503

  14. Physiological Information Database (PID)

    EPA Science Inventory

    EPA has developed a physiological information database (created using Microsoft ACCESS) intended to be used in PBPK modeling. The database contains physiological parameter values for humans from early childhood through senescence as well as similar data for laboratory animal spec...

  15. THE ECOTOX DATABASE

    EPA Science Inventory

    The database provides chemical-specific toxicity information for aquatic life, terrestrial plants, and terrestrial wildlife. ECOTOX is a comprehensive ecotoxicology database and is therefore essential for providing and suppoirting high quality models needed to estimate population...

  16. The Neotoma Paleoecology Database

    NASA Astrophysics Data System (ADS)

    Grimm, E. C.; Ashworth, A. C.; Barnosky, A. D.; Betancourt, J. L.; Bills, B.; Booth, R.; Blois, J.; Charles, D. F.; Graham, R. W.; Goring, S. J.; Hausmann, S.; Smith, A. J.; Williams, J. W.; Buckland, P.

    2015-12-01

    The Neotoma Paleoecology Database (www.neotomadb.org) is a multiproxy, open-access, relational database that includes fossil data for the past 5 million years (the late Neogene and Quaternary Periods). Modern distributional data for various organisms are also being made available for calibration and paleoecological analyses. The project is a collaborative effort among individuals from more than 20 institutions worldwide, including domain scientists representing a spectrum of Pliocene-Quaternary fossil data types, as well as experts in information technology. Working groups are active for diatoms, insects, ostracodes, pollen and plant macroscopic remains, testate amoebae, rodent middens, vertebrates, age models, geochemistry and taphonomy. Groups are also active in developing online tools for data analyses and for developing modules for teaching at different levels. A key design concept of NeotomaDB is that stewards for various data types are able to remotely upload and manage data. Cooperatives for different kinds of paleo data, or from different regions, can appoint their own stewards. Over the past year, much progress has been made on development of the steward software-interface that will enable this capability. The steward interface uses web services that provide access to the database. More generally, these web services enable remote programmatic access to the database, which both desktop and web applications can use and which provide real-time access to the most current data. Use of these services can alleviate the need to download the entire database, which can be out-of-date as soon as new data are entered. In general, the Neotoma web services deliver data either from an entire table or from the results of a view. Upon request, new web services can be quickly generated. Future developments will likely expand the spatial and temporal dimensions of the database. NeotomaDB is open to receiving new datasets and stewards from the global Quaternary community

  17. Organic acid modeling and model validation: Workshop summary

    SciTech Connect

    Sullivan, T.J.; Eilers, J.M.

    1992-08-14

    A workshop was held in Corvallis, Oregon on April 9--10, 1992 at the offices of E S Environmental Chemistry, Inc. The purpose of this workshop was to initiate research efforts on the entitled Incorporation of an organic acid representation into MAGIC (Model of Acidification of Groundwater in Catchments) and testing of the revised model using Independent data sources.'' The workshop was attended by a team of internationally-recognized experts in the fields of surface water acid-bass chemistry, organic acids, and watershed modeling. The rationale for the proposed research is based on the recent comparison between MAGIC model hindcasts and paleolimnological inferences of historical acidification for a set of 33 statistically-selected Adirondack lakes. Agreement between diatom-inferred and MAGIC-hindcast lakewater chemistry in the earlier research had been less than satisfactory. Based on preliminary analyses, it was concluded that incorporation of a reasonable organic acid representation into the version of MAGIC used for hindcasting was the logical next step toward improving model agreement.

  18. Organic acid modeling and model validation: Workshop summary. Final report

    SciTech Connect

    Sullivan, T.J.; Eilers, J.M.

    1992-08-14

    A workshop was held in Corvallis, Oregon on April 9--10, 1992 at the offices of E&S Environmental Chemistry, Inc. The purpose of this workshop was to initiate research efforts on the entitled ``Incorporation of an organic acid representation into MAGIC (Model of Acidification of Groundwater in Catchments) and testing of the revised model using Independent data sources.`` The workshop was attended by a team of internationally-recognized experts in the fields of surface water acid-bass chemistry, organic acids, and watershed modeling. The rationale for the proposed research is based on the recent comparison between MAGIC model hindcasts and paleolimnological inferences of historical acidification for a set of 33 statistically-selected Adirondack lakes. Agreement between diatom-inferred and MAGIC-hindcast lakewater chemistry in the earlier research had been less than satisfactory. Based on preliminary analyses, it was concluded that incorporation of a reasonable organic acid representation into the version of MAGIC used for hindcasting was the logical next step toward improving model agreement.

  19. JAK/STAT signalling--an executable model assembled from molecule-centred modules demonstrating a module-oriented database concept for systems and synthetic biology.

    PubMed

    Blätke, Mary Ann; Dittrich, Anna; Rohr, Christian; Heiner, Monika; Schaper, Fred; Marwan, Wolfgang

    2013-06-01

    Mathematical models of molecular networks regulating biological processes in cells or organisms are most frequently designed as sets of ordinary differential equations. Various modularisation methods have been applied to reduce the complexity of models, to analyse their structural properties, to separate biological processes, or to reuse model parts. Taking the JAK/STAT signalling pathway with the extensive combinatorial cross-talk of its components as a case study, we make a natural approach to modularisation by creating one module for each biomolecule. Each module consists of a Petri net and associated metadata and is organised in a database publically accessible through a web interface (). The Petri net describes the reaction mechanism of a given biomolecule and its functional interactions with other components including relevant conformational states. The database is designed to support the curation, documentation, version control, and update of individual modules, and to assist the user in automatically composing complex models from modules. Biomolecule centred modules, associated metadata, and database support together allow the automatic creation of models by considering differential gene expression in given cell types or under certain physiological conditions or states of disease. Modularity also facilitates exploring the consequences of alternative molecular mechanisms by comparative simulation of automatically created models even for users without mathematical skills. Models may be selectively executed as an ODE system, stochastic, or qualitative models or hybrid and exported in the SBML format. The fully automated generation of models of redesigned networks by metadata-guided modification of modules representing biomolecules with mutated function or specificity is proposed. PMID:23443149

  20. FEMA (Federal Emergency Management Agency) database requirements assessment and resource directory model. Final report 24 Aug 81-15 May 82

    SciTech Connect

    Tenopir, C.; Williams, M.E.

    1982-05-01

    Word-oriented databases (bibliographic, textual, directory, etc.) relevant to various units within the Federal Emergency Management Agency are identified and those of most potential relevance are analyzed. Subject profiles reflecting the interests of each major FEMA unit were developed and tested online on fifteen publicly available databases. The databases were then ranked by the number of citations pertinent to all aspects of emergency management and the number of pertinent citations per year of database coverage. Sample citations from the fifteen databases are included. A model Directory of Databases pertinent to emergency management was developed.

  1. Carbonatites of the World, Explored Deposits of Nb and REE - Database and Grade and Tonnage Models

    USGS Publications Warehouse

    Berger, Vladimir I.; Singer, Donald A.; Orris, Greta J.

    2009-01-01

    This report is based on published tonnage and grade data on 58 Nb- and rare-earth-element (REE)-bearing carbonatite deposits that are mostly well explored and are partially mined or contain resources of these elements. The deposits represent only a part of the known 527 carbonatites around the world, but they are characterized by reliable quantitative data on ore tonnages and grades of niobium and REE. Grade and tonnage models are an important component of mineral resource assessments. Carbonatites present one of the main natural sources of niobium and rare-earth elements, the economic importance of which grows consistently. A purpose of this report is to update earlier publications. New information about known deposits, as well as data on new deposits published during the last decade, are incorporated in the present paper. The compiled database (appendix 1; linked to right) contains 60 explored Nb- and REE-bearing carbonatite deposits - resources of 55 of these deposits are taken from publications. In the present updated grade-tonnage model we have added 24 deposits comparing with the previous model of Singer (1998). Resources of most deposits are residuum ores in the upper part of carbonatite bodies. Mineral-deposit models are important in exploration planning and quantitative resource assessments for two reasons: (1) grades and tonnages among deposit types vary significantly, and (2) deposits of different types are present in distinct geologic settings that can be identified from geologic maps. Mineral-deposit models combine the diverse geoscience information on geology, mineral occurrences, geophysics, and geochemistry used in resource assessments and mineral exploration. Globally based deposit models allow recognition of important features and demonstrate how common different features are. Well-designed deposit models allow geologists to deduce possible mineral-deposit types in a given geologic environment, and the grade and tonnage models allow economists to

  2. Towards Global QSAR Model Building for Acute Toxicity: Munro Database Case Study

    PubMed Central

    Chavan, Swapnil; Nicholls, Ian A.; Karlsson, Björn C. G.; Rosengren, Annika M.; Ballabio, Davide; Consonni, Viviana; Todeschini, Roberto

    2014-01-01

    A series of 436 Munro database chemicals were studied with respect to their corresponding experimental LD50 values to investigate the possibility of establishing a global QSAR model for acute toxicity. Dragon molecular descriptors were used for the QSAR model development and genetic algorithms were used to select descriptors better correlated with toxicity data. Toxic values were discretized in a qualitative class on the basis of the Globally Harmonized Scheme: the 436 chemicals were divided into 3 classes based on their experimental LD50 values: highly toxic, intermediate toxic and low to non-toxic. The k-nearest neighbor (k-NN) classification method was calibrated on 25 molecular descriptors and gave a non-error rate (NER) equal to 0.66 and 0.57 for internal and external prediction sets, respectively. Even if the classification performances are not optimal, the subsequent analysis of the selected descriptors and their relationship with toxicity levels constitute a step towards the development of a global QSAR model for acute toxicity. PMID:25302621

  3. Modeling charge transport in organic photovoltaic materials.

    PubMed

    Nelson, Jenny; Kwiatkowski, Joe J; Kirkpatrick, James; Frost, Jarvist M

    2009-11-17

    The performance of an organic photovoltaic cell depends critically on the mobility of charge carriers within the constituent molecular semiconductor materials. However, a complex combination of phenomena that span a range of length and time scales control charge transport in disordered organic semiconductors. As a result, it is difficult to rationalize charge transport properties in terms of material parameters. Until now, efforts to improve charge mobilities in molecular semiconductors have proceeded largely by trial and error rather than through systematic design. However, recent developments have enabled the first predictive simulation studies of charge transport in disordered organic semiconductors. This Account describes a set of computational methods, specifically molecular modeling methods, to simulate molecular packing, quantum chemical calculations of charge transfer rates, and Monte Carlo simulations of charge transport. Using case studies, we show how this combination of methods can reproduce experimental mobilities with few or no fitting parameters. Although currently applied to material systems of high symmetry or well-defined structure, further developments of this approach could address more complex systems such anisotropic or multicomponent solids and conjugated polymers. Even with an approximate treatment of packing disorder, these computational methods simulate experimental mobilities within an order of magnitude at high electric fields. We can both reproduce the relative values of electron and hole mobility in a conjugated small molecule and rationalize those values based on the symmetry of frontier orbitals. Using fully atomistic molecular dynamics simulations of molecular packing, we can quantitatively replicate vertical charge transport along stacks of discotic liquid crystals which vary only in the structure of their side chains. We can reproduce the trends in mobility with molecular weight for self-organizing polymers using a cheap, coarse

  4. Spectral Line-Shape Model to Replace the Voigt Profile in Spectroscopic Databases

    NASA Astrophysics Data System (ADS)

    Lisak, Daniel; Ngo, Ngoc Hoa; Tran, Ha; Hartmann, Jean-Michel

    2014-06-01

    The standard description of molecular line shapes in spectral databases and radiative transfer codes is based on the Voigt profile. It is well known that its simplified assumptions of absorber free motion and independence of collisional parameters from absorber velocity lead to systematic errors in analysis of experimental spectra, and retrieval of gas concentration. We demonstrate1,2 that the partially correlated quadratic speed-dependent hardcollision profile3. (pCqSDHCP) is a good candidate to replace the Voigt profile in the next generations of spectroscopic databases. This profile takes into account the following physical effects: the Doppler broadening, the pressure broadening and shifting of the line, the velocity-changing collisions, the speed-dependence of pressure broadening and shifting, and correlations between velocity- and phase/state-changing collisions. The speed-dependence of pressure broadening and shifting is incorporated into the pCqSDNGP in the so-called quadratic approximation. The velocity-changing collisions lead to the Dicke narrowing effect; however in many cases correlations between velocityand phase/state-changing collisions may lead to effective reduction of observed Dicke narrowing. The hard-collision model of velocity-changing collisions is also known as the Nelkin-Ghatak model or Rautian model. Applicability of the pCqSDHCP for different molecular systems was tested on calculated and experimental spectra of such molecules as H2, O2, CO2, H2O in a wide span of pressures. For all considered systems, pCqSDHCP is able to describe molecular spectra at least an order of magnitude better than the Voigt profile with all fitted parameters being linear with pressure. In the most cases pCqSDHCP can reproduce the reference spectra down to 0.2% or better, which fulfills the requirements of the most demanding remote-sensing applications. An important advantage of pCqSDHCP is that a fast algorithm for its computation was developedab4,5 and allows

  5. Modeling plant-level industrial energy demand with the Manufacturing Energy Consumption Survey (MECS) database and the Longitudinal Research Database (LRD)

    SciTech Connect

    Boyd, G.A.; Neifer, M.J.; Ross, M.H.

    1992-08-01

    This report discusses Phase 1 of a project to help the US Department of Energy determine the applicability of the Manufacturing Energy Consumption Survey (MECS) database and the Longitudinal Research Database (LRD) for industrial modeling and analysis. Research was conducted at the US Bureau of the Census; disclosure of the MECS/LRD data used as a basis for this report was subject to the Bureau`s confidentiality restriction. The project is designed to examine the plant-level energy behavior of energy-intensive industries. In Phase 1, six industries at the four-digit standard industrial classification (SIC) level were studied. The utility of analyzing four-digit SIC samples at the plant level is mixed, but the plant-level structure of the MECS/LRD makes analyzing samples disaggregated below the four-digit level feasible, particularly when the MECS/LRD data are combined with trade association or other external data. When external data are used, the validity of using value of shipments as a measure of output for analyzing energy use can also be examined. Phase 1 results indicate that technical efficiency and the distribution of energy intensities vary significantly at the plant level. They also show that the six industries exhibit monopsony-like behavior; that is, energy prices vary significantly at the plant level, with lower prices being correlated with a higher level of energy consumption. Finally, they show to what degree selected energy-intensive products are manufactured outside their primary industry.

  6. High prevalence of multistability of rest states and bursting in a database of a model neuron.

    PubMed

    Marin, Bóris; Barnett, William H; Doloc-Mihu, Anca; Calabrese, Ronald L; Cymbalyuk, Gennady S

    2013-01-01

    Flexibility in neuronal circuits has its roots in the dynamical richness of their neurons. Depending on their membrane properties single neurons can produce a plethora of activity regimes including silence, spiking and bursting. What is less appreciated is that these regimes can coexist with each other so that a transient stimulus can cause persistent change in the activity of a given neuron. Such multistability of the neuronal dynamics has been shown in a variety of neurons under different modulatory conditions. It can play either a functional role or present a substrate for dynamical diseases. We considered a database of an isolated leech heart interneuron model that can display silent, tonic spiking and bursting regimes. We analyzed only the cases of endogenous bursters producing functional half-center oscillators (HCOs). Using a one parameter (the leak conductance (g(leak)) bifurcation analysis, we extended the database to include silent regimes (stationary states) and systematically classified cases for the coexistence of silent and bursting regimes. We showed that different cases could exhibit two stable depolarized stationary states and two hyperpolarized stationary states in addition to various spiking and bursting regimes. We analyzed all cases of endogenous bursters and found that 18% of the cases were multistable, exhibiting coexistences of stationary states and bursting. Moreover, 91% of the cases exhibited multistability in some range of g(leak). We also explored HCOs built of multistable neuron cases with coexisting stationary states and a bursting regime. In 96% of cases analyzed, the HCOs resumed normal alternating bursting after one of the neurons was reset to a stationary state, proving themselves robust against this perturbation.

  7. Earthquake Model of the Middle East (EMME) Project: Active Fault Database for the Middle East Region

    NASA Astrophysics Data System (ADS)

    Gülen, L.; Wp2 Team

    2010-12-01

    The Earthquake Model of the Middle East (EMME) Project is a regional project of the umbrella GEM (Global Earthquake Model) project (http://www.emme-gem.org/). EMME project region includes Turkey, Georgia, Armenia, Azerbaijan, Syria, Lebanon, Jordan, Iran, Pakistan, and Afghanistan. Both EMME and SHARE projects overlap and Turkey becomes a bridge connecting the two projects. The Middle East region is tectonically and seismically very active part of the Alpine-Himalayan orogenic belt. Many major earthquakes have occurred in this region over the years causing casualties in the millions. The EMME project will use PSHA approach and the existing source models will be revised or modified by the incorporation of newly acquired data. More importantly the most distinguishing aspect of the EMME project from the previous ones will be its dynamic character. This very important characteristic is accomplished by the design of a flexible and scalable database that will permit continuous update, refinement, and analysis. A digital active fault map of the Middle East region is under construction in ArcGIS format. We are developing a database of fault parameters for active faults that are capable of generating earthquakes above a threshold magnitude of Mw≥5.5. Similar to the WGCEP-2007 and UCERF-2 projects, the EMME project database includes information on the geometry and rates of movement of faults in a “Fault Section Database”. The “Fault Section” concept has a physical significance, in that if one or more fault parameters change, a new fault section is defined along a fault zone. So far over 3,000 Fault Sections have been defined and parameterized for the Middle East region. A separate “Paleo-Sites Database” includes information on the timing and amounts of fault displacement for major fault zones. A digital reference library that includes the pdf files of the relevant papers, reports is also being prepared. Another task of the WP-2 of the EMME project is to prepare

  8. Data-based information gain on the response behaviour of hydrological models at catchment scale

    NASA Astrophysics Data System (ADS)

    Willems, Patrick

    2013-04-01

    A data-based approach is presented to analyse the response behaviour of hydrological models at the catchment scale. The approach starts with a number of sequential time series processing steps, applied to available rainfall, ETo and river flow observation series. These include separation of the high frequency (e.g., hourly, daily) river flow series into subflows, split of the series in nearly independent quick and slow flow hydrograph periods, and the extraction of nearly independent peak and low flows. Quick-, inter- and slow-subflow recession behaviour, sub-responses to rainfall and soil water storage are derived from the time series data. This data-based information on the catchment response behaviour can be applied on the basis of: - Model-structure identification and case-specific construction of lumped conceptual models for gauged catchments; or diagnostic evaluation of existing model structures; - Intercomparison of runoff responses for gauged catchments in a river basin, in order to identify similarity or significant differences between stations or between time periods, and relate these differences to spatial differences or temporal changes in catchment characteristics; - (based on the evaluation of the temporal changes in previous point:) Detection of temporal changes/trends and identification of its causes: climate trends, or land use changes; - Identification of asymptotic properties of the rainfall-runoff behaviour towards extreme peak or low flow conditions (for a given catchment) or towards extreme catchment conditions (for regionalization, ungauged basin prediction purposes); hence evaluating the performance of the model in making extrapolations beyond the range of available stations' data; - (based on the evaluation in previous point:) Evaluation of the usefulness of the model for making extrapolations to more extreme climate conditions projected by for instance climate models. Examples are provided for river basins in Belgium, Ethiopia, Kenya

  9. Making designer mutants in model organisms

    PubMed Central

    Peng, Ying; Clark, Karl J.; Campbell, Jarryd M.; Panetta, Magdalena R.; Guo, Yi; Ekker, Stephen C.

    2014-01-01

    Recent advances in the targeted modification of complex eukaryotic genomes have unlocked a new era of genome engineering. From the pioneering work using zinc-finger nucleases (ZFNs), to the advent of the versatile and specific TALEN systems, and most recently the highly accessible CRISPR/Cas9 systems, we now possess an unprecedented ability to analyze developmental processes using sophisticated designer genetic tools. In this Review, we summarize the common approaches and applications of these still-evolving tools as they are being used in the most popular model developmental systems. Excitingly, these robust and simple genomic engineering tools also promise to revolutionize developmental studies using less well established experimental organisms. PMID:25336735

  10. iGNM 2.0: the Gaussian network model database for biomolecular structural dynamics

    PubMed Central

    Li, Hongchun; Chang, Yuan-Yu; Yang, Lee-Wei; Bahar, Ivet

    2016-01-01

    Gaussian network model (GNM) is a simple yet powerful model for investigating the dynamics of proteins and their complexes. GNM analysis became a broadly used method for assessing the conformational dynamics of biomolecular structures with the development of a user-friendly interface and database, iGNM, in 2005. We present here an updated version, iGNM 2.0 http://gnmdb.csb.pitt.edu/, which covers more than 95% of the structures currently available in the Protein Data Bank (PDB). Advanced search and visualization capabilities, both 2D and 3D, permit users to retrieve information on inter-residue and inter-domain cross-correlations, cooperative modes of motion, the location of hinge sites and energy localization spots. The ability of iGNM 2.0 to provide structural dynamics data on the large majority of PDB structures and, in particular, on their biological assemblies makes it a useful resource for establishing the bridge between structure, dynamics and function. PMID:26582920

  11. Cazymes Analysis Toolkit (CAT): Webservice for searching and analyzing carbohydrateactive enzymes in a newly sequenced organism using CAZy database

    SciTech Connect

    Karpinets, Tatiana V; Park, Byung; Syed, Mustafa H; Uberbacher, Edward C; Leuze, Michael Rex

    2010-01-01

    The Carbohydrate-Active Enzyme (CAZy) database provides a rich set of manually annotated enzymes that degrade, modify, or create glycosidic bonds. Despite rich and invaluable information stored in the database, software tools utilizing this information for annotation of newly sequenced genomes by CAZy families are limited. We have employed two annotation approaches to fill the gap between manually curated high-quality protein sequences collected in the CAZy database and the growing number of other protein sequences produced by genome or metagenome sequencing projects. The first approach is based on a similarity search against the entire non-redundant sequences of the CAZy database. The second approach performs annotation using links or correspondences between the CAZy families and protein family domains. The links were discovered using the association rule learning algorithm applied to sequences from the CAZy database. The approaches complement each other and in combination achieved high specificity and sensitivity when cross-evaluated with the manually curated genomes of Clostridium thermocellum ATCC 27405 and Saccharophagus degradans 2-40. The capability of the proposed framework to predict the function of unknown protein domains (DUF) and of hypothetical proteins in the genome of Neurospora crassa is demonstrated. The framework is implemented as a Web service, the CAZymes Analysis Toolkit (CAT), and is available at http://cricket.ornl.gov/cgi-bin/cat.cgi.

  12. Data-based modelling and environmental sensitivity of vegetation in China

    NASA Astrophysics Data System (ADS)

    Wang, H.; Prentice, I. C.; Ni, J.

    2013-01-01

    A process-oriented niche specification (PONS) model was constructed to quantify climatic controls on the distribution of ecosystems, based on the vegetation map of China. PONS uses general hypotheses about bioclimatic controls to provide a "bridge" between statistical niche models and more complex process-based models. Canonical correspondence analysis provided an overview of relationships between the abundances of 55 plant communities in 0.1° grid cells and associated mean values of 20 predictor variables. Of these, GDD (accumulated degree days above 0 °C) Cramer-Prentice α (an estimate of the ratio of actual to equilibrium evapotranspiration) and mGDD5 (mean temperature during the period above 5 °C) showed the greatest predictive power. These three variables were used to develop generalized linear models for the probability of occurrence of 16 vegetation classes, aggregated from the original 55 types by k-means clustering according to bioclimatic similarity. Each class was hypothesized to possess a unimodal relationship to each bioclimate variable, independently of the other variables. A simple calibration was used to generate vegetation maps from the predicted probabilities of the classes. Modelled and observed vegetation maps showed good to excellent agreement (κ = 0.745). A sensitivity study examined modelled responses of vegetation distribution to spatially uniform changes in temperature, precipitation and [CO2], the latter included via an offset to α (based on an independent, data-based light use efficiency model for forest net primary production). Warming shifted the boundaries of most vegetation classes northward and westward while temperate steppe and desert replaced alpine tundra and steppe in the southeast of the Tibetan Plateau. Increased precipitation expanded mesic vegetation at the expense of xeric vegetation. The effect of [CO2] doubling was roughly equivalent to increasing precipitation by ∼ 30%, favouring woody vegetation types

  13. Data-based modelling and environmental sensitivity of vegetation in China

    NASA Astrophysics Data System (ADS)

    Wang, H.; Prentice, I. C.; Ni, J.

    2013-09-01

    A process-oriented niche specification (PONS) model was constructed to quantify climatic controls on the distribution of ecosystems, based on the vegetation map of China. PONS uses general hypotheses about bioclimatic controls to provide a "bridge" between statistical niche models and more complex process-based models. Canonical correspondence analysis provided an overview of relationships between the abundances of 55 plant communities in 0.1° grid cells and associated mean values of 20 predictor variables. Of these, GDD0 (accumulated degree days above 0 °C), Cramer-Prentice α (an estimate of the ratio of actual to equilibrium evapotranspiration) and mGDD5 (mean temperature during the period above 5 °C) showed the greatest predictive power. These three variables were used to develop generalized linear models for the probability of occurrence of 16 vegetation classes, aggregated from the original 55 types by k-means clustering according to bioclimatic similarity. Each class was hypothesized to possess a unimodal relationship to each bioclimate variable, independently of the other variables. A simple calibration was used to generate vegetation maps from the predicted probabilities of the classes. Modelled and observed vegetation maps showed good to excellent agreement (κ = 0.745). A sensitivity study examined modelled responses of vegetation distribution to spatially uniform changes in temperature, precipitation and [CO2], the latter included via an offset to α (based on an independent, data-based light use efficiency model for forest net primary production). Warming shifted the boundaries of most vegetation classes northward and westward while temperate steppe and desert replaced alpine tundra and steppe in the southeast of the Tibetan Plateau. Increased precipitation expanded mesic vegetation at the expense of xeric vegetation. The effect of [CO2] doubling was roughly equivalent to increasing precipitation by ~ 30%, favouring woody vegetation types

  14. AgBase: supporting functional modeling in agricultural organisms

    PubMed Central

    McCarthy, Fiona M.; Gresham, Cathy R.; Buza, Teresia J.; Chouvarine, Philippe; Pillai, Lakshmi R.; Kumar, Ranjit; Ozkan, Seval; Wang, Hui; Manda, Prashanti; Arick, Tony; Bridges, Susan M.; Burgess, Shane C.

    2011-01-01

    AgBase (http://www.agbase.msstate.edu/) provides resources to facilitate modeling of functional genomics data and structural and functional annotation of agriculturally important animal, plant, microbe and parasite genomes. The website is redesigned to improve accessibility and ease of use, including improved search capabilities. Expanded capabilities include new dedicated pages for horse, cat, dog, cotton, rice and soybean. We currently provide 590 240 Gene Ontology (GO) annotations to 105 454 gene products in 64 different species, including GO annotations linked to transcripts represented on agricultural microarrays. For many of these arrays, this provides the only functional annotation available. GO annotations are available for download and we provide comprehensive, species-specific GO annotation files for 18 different organisms. The tools available at AgBase have been expanded and several existing tools improved based upon user feedback. One of seven new tools available at AgBase, GOModeler, supports hypothesis testing from functional genomics data. We host several associated databases and provide genome browsers for three agricultural pathogens. Moreover, we provide comprehensive training resources (including worked examples and tutorials) via links to Educational Resources at the AgBase website. PMID:21075795

  15. Evaluation of the quality indices of landslide databases for their use in susceptibility models

    NASA Astrophysics Data System (ADS)

    Iadanza, C.; Trigila, A.; Spizzichino, D.

    2009-04-01

    The aim of this study is to define quality indices for evaluation of the homogeneity, completeness and accuracy of landslide databases. This evaluation is absolutely essential if the databases are to be used for the definition and implementation of predictive models for estimation of landslide susceptibility. In particular, descriptions are given of the quality indices defined for the Italian Landslide Inventory (IFFI Project). The IFFI Project, with the funding of 4.1 million Euro in 1997 by the Italian Government, is implemented by ISPRA (Institute for Environmental Protection and Research/Geological Survey of Italy) and by the Regions and the Self-Governing Provinces. The Project collects and maps the landslides over the entire Italian territory in accordance with standardised and shared methods. A specific "Landslide Data Sheet" has been prepared in order to collect all the landslide information, subdivided into three levels of progressively increasing detail. A detailed landslide mapping (scale 1:10,000) has been adopted throughout most of Italy. A scale of 1:25,000 has been used in high mountainous areas or in low population areas. With regard to the mapping, every landslide is represented by a georeferenced point placed, by convention, at the highest point of the landslide crown; a polygon, if the landslide may be mapped at the adopted survey scale; a line when the landslides have a very elongated form with a width which may not be mapped. The Inventory contains 482,272 landslides covering an area of approximately 20,500 km2, which is equivalent to 6.8% of the Italian territory. With regard to the homogeneity and completeness indices, comparisons have been made of the landslide indices for areas with the same outcropping lithology, statistical analyses of the landslide area have been carried out and information from the IFFI Inventory has been compared with that contained in other databases (e.g. AVI - Inventory of information on sites historically affected

  16. MEROPS: the peptidase database.

    PubMed

    Rawlings, N D; Barrett, A J

    1999-01-01

    The MEROPS database (http://www.bi.bbsrc.ac.uk/Merops/Merops.+ ++htm) provides a catalogue and structure-based classification of peptidases (i.e. all proteolytic enzymes). This is a large group of proteins (approximately 2% of all gene products) that is of particular importance in medicine and biotechnology. An index of the peptidases by name or synonym gives access to a set of files termed PepCards each of which provides information on a single peptidase. Each card file contains information on classification and nomenclature, and hypertext links to the relevant entries in online databases for human genetics, protein and nucleic acid sequence data and tertiary structure. Another index provides access to the PepCards by organism name so that the user can retrieve all known peptidases from a particular species. The peptidases are classified into families on the basis of statistically significant similarities between the protein sequences in the part termed the 'peptidase unit' that is most directly responsible for activity. Families that are thought to have common evolutionary origins and are known or expected to have similar tertiary folds are grouped into clans. The MEROPS database provides sets of files called FamCards and ClanCards describing the individual families and clans. Each FamCard document provides links to other databases for sequence motifs and secondary and tertiary structures, and shows the distribution of the family across the major kingdoms of living creatures. Release 3.03 of MEROPS contains 758 peptidases, 153 families and 22 clans. We suggest that the MEROPS database provides a model for a way in which a system of classification for a functional group of proteins can be developed and used as an organizational framework around which to assemble a variety of related information.

  17. Models of care and organization of services.

    PubMed

    Markova, Alina; Xiong, Michael; Lester, Jenna; Burnside, Nancy J

    2012-01-01

    This article examines the overall organization of services and delivery of health care in the United States. Health maintenance organization, fee-for-service, preferred provider organizations, and the Veterans Health Administration are discussed, with a focus on structure, outcomes, and areas for improvement. An overview of wait times, malpractice, telemedicine, and the growing population of physician extenders in dermatology is also provided.

  18. FACILITATING ADVANCED URBAN METEOROLOGY AND AIR QUALITY MODELING CAPABILITIES WITH HIGH RESOLUTION URBAN DATABASE AND ACCESS PORTAL TOOLS

    EPA Science Inventory

    Information of urban morphological features at high resolution is needed to properly model and characterize the meteorological and air quality fields in urban areas. We describe a new project called National Urban Database with Access Portal Tool, (NUDAPT) that addresses this nee...

  19. Statistical model for large-scale peptide identification in databases from tandem mass spectra using SEQUEST.

    PubMed

    López-Ferrer, Daniel; Martínez-Bartolomé, Salvador; Villar, Margarita; Campillos, Mónica; Martín-Maroto, Fernando; Vázquez, Jesús

    2004-12-01

    Recent technological advances have made multidimensional peptide separation techniques coupled with tandem mass spectrometry the method of choice for high-throughput identification of proteins. Due to these advances, the development of software tools for large-scale, fully automated, unambiguous peptide identification is highly necessary. In this work, we have used as a model the nuclear proteome from Jurkat cells and present a processing algorithm that allows accurate predictions of random matching distributions, based on the two SEQUEST scores Xcorr and DeltaCn. Our method permits a very simple and precise calculation of the probabilities associated with individual peptide assignments, as well as of the false discovery rate among the peptides identified in any experiment. A further mathematical analysis demonstrates that the score distributions are highly dependent on database size and precursor mass window and suggests that the probability associated with SEQUEST scores depends on the number of candidate peptide sequences available for the search. Our results highlight the importance of adjusting the filtering criteria to discriminate between correct and incorrect peptide sequences according to the circumstances of each particular experiment.

  20. Anatomical database generation for radiation transport modeling from computed tomography (CT) scan data

    SciTech Connect

    Margle, S.M.; Tinnel, E.P.; Till, L.E.; Eckerman, K.F.; Durfee, R.C.

    1989-01-01

    Geometric models of the anatomy are used routinely in calculations of the radiation dose in organs and tissues of the body. Development of such models has been hampered by lack of detailed anatomical information on children, and models themselves have been limited to quadratic conic sections. This summary reviews the development of an image processing workstation used to extract anatomical information from routine diagnostic CT procedure. A standard IBM PC/AT microcomputer has been augmented with an automatically loading 9-track magnetic tape drive, an 8-bit 1024 {times} 1024 pixel graphics adapter/monitor/film recording package, a mouse/trackball assembly, dual 20 MB removable cartridge media, a 72 MB disk drive, and a printer. Software utilized by the workstation includes a Geographic Information System (modified for manipulation of CT images), CAD software, imaging software, and various modules to ease data transfer among the software packages. 5 refs., 3 figs.

  1. Sediment-hosted gold deposits of the world: database and grade and tonnage models

    USGS Publications Warehouse

    Berger, Vladimir I.; Mosier, Dan L.; Bliss, James D.; Moring, Barry C.

    2014-01-01

    All sediment-hosted gold deposits (as a single population) share one characteristic—they all have disseminated micron-sized invisible gold in sedimentary rocks. Sediment-hosted gold deposits are recognized in the Great Basin province of the western United States and in China along with a few recognized deposits in Indonesia, Iran, and Malaysia. Three new grade and tonnage models for sediment-hosted gold deposits are presented in this paper: (1) a general sediment-hosted gold type model, (2) a Carlin subtype model, and (3) a Chinese subtype model. These models are based on grade and tonnage data from a database compilation of 118 sediment-hosted gold deposits including a total of 123 global deposits. The new general grade and tonnage model for sediment-hosted gold deposits (n=118) has a median tonnage of 5.7 million metric tonnes (Mt) and a gold grade of 2.9 grams per tonne (g/t). This new grade and tonnage model is remarkable in that the estimated parameters of the resulting grade and tonnage distributions are comparable to the previous model of Mosier and others (1992). A notable change is in the reporting of silver in more than 10 percent of deposits; moreover, the previous model had not considered deposits in China. From this general grade and tonnage model, two significantly different subtypes of sediment-hosted gold deposits are differentiated: Carlin and Chinese. The Carlin subtype includes 88 deposits in the western United States, Indonesia, Iran, and Malaysia, with median tonnage and grade of 7.1 Mt and 2.0 g/t Au, respectively. The silver grade is 0.78 g/t Ag for the 10th percentile of deposits. The Chinese subtype represents 30 deposits in China, with a median tonnage of 3.9 Mt and medium grade of 4.6 g/t Au. Important differences are recognized in the mineralogy and alteration of the two sediment-hosted gold subtypes such as: increased sulfide minerals in the Chinese subtype and decalcification alteration dominant in the Carlin type. We therefore

  2. The Time Is Right to Focus on Model Organism Metabolomes.

    PubMed

    Edison, Arthur S; Hall, Robert D; Junot, Christophe; Karp, Peter D; Kurland, Irwin J; Mistrik, Robert; Reed, Laura K; Saito, Kazuki; Salek, Reza M; Steinbeck, Christoph; Sumner, Lloyd W; Viant, Mark R

    2016-02-15

    Model organisms are an essential component of biological and biomedical research that can be used to study specific biological processes. These organisms are in part selected for facile experimental study. However, just as importantly, intensive study of a small number of model organisms yields important synergies as discoveries in one area of science for a given organism shed light on biological processes in other areas, even for other organisms. Furthermore, the extensive knowledge bases compiled for each model organism enable systems-level understandings of these species, which enhance the overall biological and biomedical knowledge for all organisms, including humans. Building upon extensive genomics research, we argue that the time is now right to focus intensively on model organism metabolomes. We propose a grand challenge for metabolomics studies of model organisms: to identify and map all metabolites onto metabolic pathways, to develop quantitative metabolic models for model organisms, and to relate organism metabolic pathways within the context of evolutionary metabolomics, i.e., phylometabolomics. These efforts should focus on a series of established model organisms in microbial, animal and plant research.

  3. The Time Is Right to Focus on Model Organism Metabolomes

    PubMed Central

    Edison, Arthur S.; Hall, Robert D.; Junot, Christophe; Karp, Peter D.; Kurland, Irwin J.; Mistrik, Robert; Reed, Laura K.; Saito, Kazuki; Salek, Reza M.; Steinbeck, Christoph; Sumner, Lloyd W.; Viant, Mark R.

    2016-01-01

    Model organisms are an essential component of biological and biomedical research that can be used to study specific biological processes. These organisms are in part selected for facile experimental study. However, just as importantly, intensive study of a small number of model organisms yields important synergies as discoveries in one area of science for a given organism shed light on biological processes in other areas, even for other organisms. Furthermore, the extensive knowledge bases compiled for each model organism enable systems-level understandings of these species, which enhance the overall biological and biomedical knowledge for all organisms, including humans. Building upon extensive genomics research, we argue that the time is now right to focus intensively on model organism metabolomes. We propose a grand challenge for metabolomics studies of model organisms: to identify and map all metabolites onto metabolic pathways, to develop quantitative metabolic models for model organisms, and to relate organism metabolic pathways within the context of evolutionary metabolomics, i.e., phylometabolomics. These efforts should focus on a series of established model organisms in microbial, animal and plant research. PMID:26891337

  4. Database and Library Development of Organic Species using Gas Chromatography and Mass Spectral Measurements in Support of the Mars Science Laboratory

    NASA Astrophysics Data System (ADS)

    Garcia, Raul; Mahaffy, Paul; Misra, Prabhakar

    2010-02-01

    Our work involves the development of an organic contaminants database that will allow us to determine which compounds are found here on Earth and would be inadvertently detected in the Mars soil and gaseous samples as impurities. It will be used for the Sample Analysis at Mars (SAM) instrumentation analysis in the Mars Science Laboratory (MSL) rover scheduled for launch in 2011. In order to develop a comprehensive target database, we utilize the NIST Mass Spectral Library, Automated Mass Spectral Deconvolution and Identification System (AMDIS) and Ion Fingerprint Deconvolution (IFD) software to analyze the GC-MS data. We have analyzed data from commercial samples, such as paint and polymers, which have not been implemented into the rover and are now analyzing actual data from pyrolyzation on the rover. We have successfully developed an initial target compound database that will aid SAM in determining whether the components being analyzed come from Mars or are contaminants from either the rover itself or the Earth environment and are continuing to make improvements and adding data to the target contaminants database. )

  5. Making the procedure manual come alive: A prototype relational database and dynamic website model for the management of nursing information.

    PubMed

    Peace, Jane; Brennan, Patricia Flatley

    2006-01-01

    The nursing procedural manual is an essential resource for clinical practice, yet insuring its currency and availability at the point of care remains an unresolved information management challenge for nurses. While standard HTML-based web pages offer significant advantage over paper compilations, employing emerging computer science tools offers even greater promise. This paper reports on the creation of a prototypical dynamic web-based nursing procedure manual driven by a relational database. We created a relational database in MySQL to manage, store, and link the procedure information, and developed PHP files to guide content retrieval, content management, and display on demand in browser-viewable format. This database driven dynamic website model is an important innovation to meet the challenge of content management and dissemination of nursing information.

  6. A Database for Propagation Models and Conversion to C++ Programming Language

    NASA Technical Reports Server (NTRS)

    Kantak, Anil V.; Angkasa, Krisjani; Rucker, James

    1996-01-01

    The telecommunications system design engineer generally needs the quantification of effects of the propagation medium (definition of the propagation channel) to design an optimal communications system. To obtain the definition of the channel, the systems engineer generally has a few choices. A search of the relevant publications such as the IEEE Transactions, CCIR's, NASA propagation handbook, etc., may be conducted to find the desired channel values. This method may need excessive amounts of time and effort on the systems engineer's part and there is a possibility that the search may not even yield the needed results. To help the researcher and the systems engineers, it was recommended by the conference participants of NASA Propagation Experimenters (NAPEX) XV (London, Ontario, Canada, June 28 and 29, 1991) that a software should be produced that would contain propagation models and the necessary prediction methods of most propagation phenomena. Moreover, the software should be flexible enough for the user to make slight changes to the models without expending a substantial effort in programming. In the past few years, a software was produced to fit these requirements as best as could be done. The software was distributed to all NAPEX participants for evaluation and use, the participant reactions, suggestions etc., were gathered and were used to improve the subsequent releases of the software. The existing database program is in the Microsoft Excel application software and works fine within the guidelines of that environment, however, recently there have been some questions about the robustness and survivability of the Excel software in the ever changing (hopefully improving) world of software packages.

  7. Toxico-Cheminformatics and QSPR Modeling of the Carcinogenic Potency Database

    EPA Science Inventory

    Report on the development of a tiered, confirmatory scheme for prediction of chemical carcinogenicity based on QSAR studies of compounds with available mutagenic and carcinogenic data. For 693 such compounds from the Carcinogenic Potency Database characterized molecular topologic...

  8. Protein-based virtual screening of chemical databases. II. Are homology models of G-Protein Coupled Receptors suitable targets?

    PubMed

    Bissantz, Caterina; Bernard, Philippe; Hibert, Marcel; Rognan, Didier

    2003-01-01

    The aim of the current study is to investigate whether homology models of G-Protein-Coupled Receptors (GPCRs) that are based on bovine rhodopsin are reliable enough to be used for virtual screening of chemical databases. Starting from the recently described 2.8 A-resolution X-ray structure of bovine rhodopsin, homology models of an "antagonist-bound" form of three human GPCRs (dopamine D3 receptor, muscarinic M1 receptor, vasopressin V1a receptor) were constructed. The homology models were used to screen three-dimensional databases using three different docking programs (Dock, FlexX, Gold) in combination with seven scoring functions (ChemScore, Dock, FlexX, Fresno, Gold, Pmf, Score). Rhodopsin-based homology models turned out to be suitable, indeed, for virtual screening since known antagonists seeded in the test databases could be distinguished from randomly chosen molecules. However, such models are not accurate enough for retrieving known agonists. To generate receptor models better suited for agonist screening, we developed a new knowledge- and pharmacophore-based modeling procedure that might partly simulate the conformational changes occurring in the active site during receptor activation. Receptor coordinates generated by this new procedure are now suitable for agonist screening. We thus propose two alternative strategies for the virtual screening of GPCR ligands, relying on a different set of receptor coordinates (antagonist-bound and agonist-bound states).

  9. Universal organic solvent-water partition coefficient model

    PubMed

    Torrens

    2000-03-01

    A method that permits a semiquantitative estimate of the partitioning of any solute between any two media is presented. As an example, the organic solvent-water partition coefficients P are calculated. Program GSCAP is written as a version of Pascal's SCAP program. The only needed parameters are the dielectric constant and molecular volume of the organic solvent. The log P results are compared with the Pomona database. The average absolute deviation is 1.48 log units and the standard deviation is 1.66 log units.

  10. Modelling the fate of persistent organic pollutants in europe: parameterization of a gridded distribution model

    SciTech Connect

    Prevedouros, Konstantinos; MacLeod, Matthew; Jones, Kevin C.; Sweetman Andrew J.

    2003-12-01

    A regionally segmented multimedia fate model for the European continent is described together with an illustrative steady-state case study examining the fate of small gamma, Greek-HCH (lindane) based on 1998 emission data. The study builds on the regionally segmented BETR North America model structure and describes the regional segmentation and parameterization for Europe. The European continent is described by a 5 degree x 5 degree grid, leading to 50 regions together with four perimetric boxes representing regions buffering the European environment. Each zone comprises seven compartments including; upper and lower atmosphere, soil, vegetation, fresh water and sediment and coastal water. Inter-regions flows of air and water are described, exploiting information originating from GIS databases and other georeferenced data. The model is primarily designed to describe the fate of Persistent Organic Pollutants (POPs) within the European environment by examining chemical partitioning and de gradation in each region, and inter-region transport either under steady-state conditions or fully dynamically. A test case scenario is presented which examines the fate of estimated spatially resolved atmospheric emissions of lindane throughout Europe within the lower atmosphere and surface soil compartments. In accordance with the predominant wind direction in Europe, the model predicts high concentrations close to the major sources as well as towards Central and Northeast regions. Elevated soil concentrations in Scandinavian soils provide further evidence of the potential of increased scavenging by forests and subsequent accumulation by organic-rich terrestrial surfaces. Initial model predictions have revealed a factor of 5 10 underestimation of lindane concentrations in the atmosphere. This is explained by an underestimation of source strength and/or an underestimation of European background levels. The model presented can further be used to predict deposition fluxes and chemical

  11. Data Model and Relational Database Design for Highway Runoff Water-Quality Metadata

    USGS Publications Warehouse

    Granato, Gregory E.; Tessler, Steven

    2001-01-01

    A National highway and urban runoff waterquality metadatabase was developed by the U.S. Geological Survey in cooperation with the Federal Highway Administration as part of the National Highway Runoff Water-Quality Data and Methodology Synthesis (NDAMS). The database was designed to catalog available literature and to document results of the synthesis in a format that would facilitate current and future research on highway and urban runoff. This report documents the design and implementation of the NDAMS relational database, which was designed to provide a catalog of available information and the results of an assessment of the available data. All the citations and the metadata collected during the review process are presented in a stratified metadatabase that contains citations for relevant publications, abstracts (or previa), and reportreview metadata for a sample of selected reports that document results of runoff quality investigations. The database is referred to as a metadatabase because it contains information about available data sets rather than a record of the original data. The database contains the metadata needed to evaluate and characterize how valid, current, complete, comparable, and technically defensible published and available information may be when evaluated for application to the different dataquality objectives as defined by decision makers. This database is a relational database, in that all information is ultimately linked to a given citation in the catalog of available reports. The main database file contains 86 tables consisting of 29 data tables, 11 association tables, and 46 domain tables. The data tables all link to a particular citation, and each data table is focused on one aspect of the information collected in the literature search and the evaluation of available information. This database is implemented in the Microsoft (MS) Access database software because it is widely used within and outside of government and is familiar to many

  12. System and method employing a self-organizing map load feature database to identify electric load types of different electric loads

    SciTech Connect

    Lu, Bin; Harley, Ronald G.; Du, Liang; Yang, Yi; Sharma, Santosh K.; Zambare, Prachi; Madane, Mayura A.

    2014-06-17

    A method identifies electric load types of a plurality of different electric loads. The method includes providing a self-organizing map load feature database of a plurality of different electric load types and a plurality of neurons, each of the load types corresponding to a number of the neurons; employing a weight vector for each of the neurons; sensing a voltage signal and a current signal for each of the loads; determining a load feature vector including at least four different load features from the sensed voltage signal and the sensed current signal for a corresponding one of the loads; and identifying by a processor one of the load types by relating the load feature vector to the neurons of the database by identifying the weight vector of one of the neurons corresponding to the one of the load types that is a minimal distance to the load feature vector.

  13. Organizations, Environments, and Models of Public Relations.

    ERIC Educational Resources Information Center

    Grunig, James E.

    Noting that little theory has been developed to explain how and why organizations choose to manage public relations, this paper argues that theorists cannot improve the practice of public relations until they can explain what public relations is and what it contributes to the functions of an organization. The paper addresses that issue by…

  14. A computational platform to maintain and migrate manual functional annotations for BioCyc databases

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Model organism databases are an important resource for information on biological pathways and genomic data. Such databases represent the accumulation of biological data, some of which has been manually curated from literature. An essential feature of these databases is the continuing data integratio...

  15. TREATABILITY DATABASE DESCRIPTION

    EPA Science Inventory

    The Drinking Water Treatability Database (TDB) presents referenced information on the control of contaminants in drinking water. It allows drinking water utilities, first responders to spills or emergencies, treatment process designers, research organizations, academics, regulato...

  16. THE CTEPP DATABASE

    EPA Science Inventory

    The CTEPP (Children's Total Exposure to Persistent Pesticides and Other Persistent Organic Pollutants) database contains a wealth of data on children's aggregate exposures to pollutants in their everyday surroundings. Chemical analysis data for the environmental media and ques...

  17. [Organ trade versus reciprocity model. An ethical evaluation].

    PubMed

    Illies, C; Weber, F

    2004-02-01

    We perform an ethical evaluation of two models that promise to solve the increasing shortage of organs for transplantations: firstly, the legalization of organ trade, and, secondly, the so called "Reciprocity Model". Thereby unrestricted respect for the individual human being serves as the ethical standard. We conclude that the Reciprocity Model is ethically much more acceptable that organ trade even if this trade were limited to Europe. In addition, the Reciprocity Model can easily be integrated into the current Eurotransplant system of organ allocation. PMID:14750056

  18. Improved AIOMFAC model parameterisation of the temperature dependence of activity coefficients for aqueous organic mixtures

    NASA Astrophysics Data System (ADS)

    Ganbavale, G.; Zuend, A.; Marcolli, C.; Peter, T.

    2014-06-01

    This study presents a new, improved parameterisation of the temperature dependence of activity coefficients in the AIOMFAC (Aerosol Inorganic-Organic Mixtures Functional groups Activity Coefficients) model applicable for aqueous as well as water-free organic solutions. For electrolyte-free organic and organic-water mixtures the AIOMFAC model uses a group-contribution approach based on UNIFAC (UNIversal quasi-chemical Functional-group Activity Coefficients). This group-contribution approach explicitly accounts for interactions among organic functional groups and between organic functional groups and water. The previous AIOMFAC version uses a simple parameterisation of the temperature dependence of activity coefficients, aimed to be applicable in the temperature range from ~275 to ~400 K. With the goal to improve the description of a wide variety of organic compounds found in atmospheric aerosols, we extend the AIOMFAC parameterisation for the functional groups carboxyl, hydroxyl, ketone, aldehyde, ether, ester, alkyl, aromatic carbon-alcohol, and aromatic hydrocarbon to atmospherically relevant low temperatures with the introduction of a new temperature dependence parameterisation. The improved temperature dependence parameterisation is derived from classical thermodynamic theory by describing effects from changes in molar enthalpy and heat capacity of a multicomponent system. Thermodynamic equilibrium data of aqueous organic and water-free organic mixtures from the literature are carefully assessed and complemented with new measurements to establish a comprehensive database, covering a wide temperature range (~190 to ~440 K) for many of the functional group combinations considered. Different experimental data types and their processing for the estimation of AIOMFAC model parameters are discussed. The new AIOMFAC parameterisation for the temperature dependence of activity coefficients from low to high temperatures shows an overall improvement of 25% in comparison to

  19. A Modeling Exercise for the Organic Classroom

    ERIC Educational Resources Information Center

    Whitlock, Christine R.

    2010-01-01

    An in-class molecular modeling exercise is described. Groups of students are given molecular models to investigate and questions about the models to answer. This exercise is a quick and effective way to review nomenclature, stereochemistry, and conformational analysis.

  20. In praise of other model organisms.

    PubMed

    Warren, Graham

    2015-02-16

    The early cell biological literature is the resting place of false starts and lost opportunities. Though replete with multiple studies of diverse organisms, a few of which served as foundations for several fields, most were not pursued, abandoned largely for technical reasons that are no longer limiting. The time has come to revisit the old literature and to resurrect the organisms that are buried there, both to uncover new mechanisms and to marvel at the richness of the cellular world.

  1. Compilation of a database on the composition of anthropogenic VOC emissions for atmospheric modeling in Europe

    NASA Astrophysics Data System (ADS)

    Theloke, J.; Friedrich, R.

    To analyse and generate air pollution control strategies and policies, e.g. efficient abatement strategies or action plans that lead to a fulfilment of air quality aims, atmospheric dispersion models (CTMs) have to be used. These models include a chemical model, where the numerous volatile organic compounds (VOCs) species are lumped together in classes. On the other hand, emission inventories usually report only total non-methane VOC (NMVOC), but not a subdivision into these classes. Thus, VOC species profiles are needed that resolve total NMVOC emission data. The objective of this publication is to present the results of a compilation of VOC species profiles that dissolve total VOC into single-species profiles for all relevant anthropogenic emission source categories and the European situation. As in atmospheric dispersion models usually modules for generating biogenic emissions are directly included, only anthropogenic emissions are addressed. VOC species profiles for 87 emission source categories have been developed. The underlying data base can be used to generate the data for all chemical mechanisms. The species profiles have been generated using recent measurements and studies on VOC species resolution and thus represent the current state of knowledge in this area. The results can be used to create input data for atmospheric dispersion models in Europe. The profiles, especially those for solvent use, still show large uncertainties. There is still an enormous need for further measurements to achieve an improved species resolution. In addition, the solvent use directive and the DECOPAINT directive of the European Commission will result in a change of the composition of paints; more water-based and high-solid paints will be used; thus the species resolution will change drastically in the next years. Of course, the species resolution for combustion and production processes also requires further improvement.

  2. Application of Large-Scale Database-Based Online Modeling to Plant State Long-Term Estimation

    NASA Astrophysics Data System (ADS)

    Ogawa, Masatoshi; Ogai, Harutoshi

    Recently, attention has been drawn to the local modeling techniques of a new idea called “Just-In-Time (JIT) modeling”. To apply “JIT modeling” to a large amount of database online, “Large-scale database-based Online Modeling (LOM)” has been proposed. LOM is a technique that makes the retrieval of neighboring data more efficient by using both “stepwise selection” and quantization. In order to predict the long-term state of the plant without using future data of manipulated variables, an Extended Sequential Prediction method of LOM (ESP-LOM) has been proposed. In this paper, the LOM and the ESP-LOM are introduced.

  3. International Comparisions Database

    National Institute of Standards and Technology Data Gateway

    International Comparisions Database (Web, free access)   The International Comparisons Database (ICDB) serves the U.S. and the Inter-American System of Metrology (SIM) with information based on Appendices B (International Comparisons), C (Calibration and Measurement Capabilities) and D (List of Participating Countries) of the Comit� International des Poids et Mesures (CIPM) Mutual Recognition Arrangement (MRA). The official source of the data is The BIPM key comparison database. The ICDB provides access to results of comparisons of measurements and standards organized by the consultative committees of the CIPM and the Regional Metrology Organizations.

  4. A Mentor/Research Model to Teach Library Skills: An Introduction to Database Searching.

    ERIC Educational Resources Information Center

    Nissen, Karen R.; Ross, Barbara A.

    1996-01-01

    As a way to teach database searching skills, students in a nursing research class were paired with, and performed literature searches for, faculty members or community health care professionals. Because students did not formulate questions or initial search terms, they concentrated on the search process, learning how professionals find and use…

  5. Application of Knowledge Discovery in Databases Methodologies for Predictive Models for Pregnancy Adverse Events

    ERIC Educational Resources Information Center

    Taft, Laritza M.

    2010-01-01

    In its report "To Err is Human", The Institute of Medicine recommended the implementation of internal and external voluntary and mandatory automatic reporting systems to increase detection of adverse events. Knowledge Discovery in Databases (KDD) allows the detection of patterns and trends that would be hidden or less detectable if analyzed by…

  6. The Cardiac Atlas Project—an imaging database for computational modeling and statistical atlases of the heart

    PubMed Central

    Fonseca, Carissa G.; Backhaus, Michael; Bluemke, David A.; Britten, Randall D.; Chung, Jae Do; Cowan, Brett R.; Dinov, Ivo D.; Finn, J. Paul; Hunter, Peter J.; Kadish, Alan H.; Lee, Daniel C.; Lima, Joao A. C.; Medrano−Gracia, Pau; Shivkumar, Kalyanam; Suinesiaputra, Avan; Tao, Wenchao; Young, Alistair A.

    2011-01-01

    Motivation: Integrative mathematical and statistical models of cardiac anatomy and physiology can play a vital role in understanding cardiac disease phenotype and planning therapeutic strategies. However, the accuracy and predictive power of such models is dependent upon the breadth and depth of noninvasive imaging datasets. The Cardiac Atlas Project (CAP) has established a large-scale database of cardiac imaging examinations and associated clinical data in order to develop a shareable, web-accessible, structural and functional atlas of the normal and pathological heart for clinical, research and educational purposes. A goal of CAP is to facilitate collaborative statistical analysis of regional heart shape and wall motion and characterize cardiac function among and within population groups. Results: Three main open-source software components were developed: (i) a database with web-interface; (ii) a modeling client for 3D + time visualization and parametric description of shape and motion; and (iii) open data formats for semantic characterization of models and annotations. The database was implemented using a three-tier architecture utilizing MySQL, JBoss and Dcm4chee, in compliance with the DICOM standard to provide compatibility with existing clinical networks and devices. Parts of Dcm4chee were extended to access image specific attributes as search parameters. To date, approximately 3000 de-identified cardiac imaging examinations are available in the database. All software components developed by the CAP are open source and are freely available under the Mozilla Public License Version 1.1 (http://www.mozilla.org/MPL/MPL-1.1.txt). Availability: http://www.cardiacatlas.org Contact: a.young@auckland.ac.nz Supplementary information: Supplementary data are available at Bioinformatics online. PMID:21737439

  7. Biofuel Database

    National Institute of Standards and Technology Data Gateway

    Biofuel Database (Web, free access)   This database brings together structural, biological, and thermodynamic data for enzymes that are either in current use or are being considered for use in the production of biofuels.

  8. Evaluating Organic Aerosol Model Performance: Impact of two Embedded Assumptions

    NASA Astrophysics Data System (ADS)

    Jiang, W.; Giroux, E.; Roth, H.; Yin, D.

    2004-05-01

    Organic aerosols are important due to their abundance in the polluted lower atmosphere and their impact on human health and vegetation. However, modeling organic aerosols is a very challenging task because of the complexity of aerosol composition, structure, and formation processes. Assumptions and their associated uncertainties in both models and measurement data make model performance evaluation a truly demanding job. Although some assumptions are obvious, others are hidden and embedded, and can significantly impact modeling results, possibly even changing conclusions about model performance. This paper focuses on analyzing the impact of two embedded assumptions on evaluation of organic aerosol model performance. One assumption is about the enthalpy of vaporization widely used in various secondary organic aerosol (SOA) algorithms. The other is about the conversion factor used to obtain ambient organic aerosol concentrations from measured organic carbon. These two assumptions reflect uncertainties in the model and in the ambient measurement data, respectively. For illustration purposes, various choices of the assumed values are implemented in the evaluation process for an air quality model based on CMAQ (the Community Multiscale Air Quality Model). Model simulations are conducted for the Lower Fraser Valley covering Southwest British Columbia, Canada, and Northwest Washington, United States, for a historical pollution episode in 1993. To understand the impact of the assumed enthalpy of vaporization on modeling results, its impact on instantaneous organic aerosol yields (IAY) through partitioning coefficients is analysed first. The analysis shows that utilizing different enthalpy of vaporization values causes changes in the shapes of IAY curves and in the response of SOA formation capability of reactive organic gases to temperature variations. These changes are then carried into the air quality model and cause substantial changes in the organic aerosol modeling

  9. An Assessment of Thermodynamic Database Effects on Reactive Transport Models' Predictions of Permeability Fields: Insights from CO2/Brine Experiments

    NASA Astrophysics Data System (ADS)

    Tutolo, B. M.; Seyfried, W. E.; Saar, M. O.

    2011-12-01

    Numerical modeling software such as TOUGHREACT, ECLIPSE, and Geochemist's Workbench provide modules that couple mineral reactive chemistry with porosity and permeability modifications to predict the behavior of energy byproducts, such as carbon dioxide, in the subsurface. Modern coders have already included increasingly complex equations that describe natural systems (i.e. mineral dissolution/precipitation kinetic parameters and porosity/permeability functions) into these and other software applications. Generally, these computer models use the bulk volumetric changes predicted by geochemical calculations to infer porosity changes, and subsequently use highly simplified porosity/permeability correlation functions, such as the Carman-Kozeny equation, to modify permeability fields. In spite of the computational complexity provided in these models, they require, as a foundation, fundamental information on the thermodynamic stability of minerals and aqueous species at a wide range of temperatures and pressures to produce accurate predictions of the geochemistry of long-term energy byproduct storage in the subsurface, even in the simplest cases. With improvements in geochemical thermodynamic databases, researchers may begin to produce more realistic simulations of the complex interactions between fluid and heat flow and geological systems. Unfortunately, the requisite thermodynamic data is often lacking, or inaccurate. In this study, therefore, we provide a discussion of geochemical thermodynamic databases, discuss the synthesis and reconciliation of the databases used in this study, and compare predictions from reactive transport software with single phase CO2/brine experiments performed at temperatures and pressures applicable to geologic storage conditions.

  10. Scaling laws and model of words organization in spoken and written language

    NASA Astrophysics Data System (ADS)

    Bian, Chunhua; Lin, Ruokuang; Zhang, Xiaoyu; Ma, Qianli D. Y.; Ivanov, Plamen Ch.

    2016-01-01

    A broad range of complex physical and biological systems exhibits scaling laws. The human language is a complex system of words organization. Studies of written texts have revealed intriguing scaling laws that characterize the frequency of words occurrence, rank of words, and growth in the number of distinct words with text length. While studies have predominantly focused on the language system in its written form, such as books, little attention is given to the structure of spoken language. Here we investigate a database of spoken language transcripts and written texts, and we uncover that words organization in both spoken language and written texts exhibits scaling laws, although with different crossover regimes and scaling exponents. We propose a model that provides insight into words organization in spoken language and written texts, and successfully accounts for all scaling laws empirically observed in both language forms.

  11. A Model for Implementing E-Learning in Iranian Organizations

    ERIC Educational Resources Information Center

    Ghaeni, Emad; Abdehagh, Babak

    2010-01-01

    This article reviews the current status of information and communications technology (ICT) usage and provides a comprehensive outlook on e-learning in both virtual universities and organizations in Iran. A model for e-learning implementation is presented. This model tries to address specific issues in Iranian organizations. (Contains 1 table and 2…

  12. Modeling the Explicit Chemistry of Anthropogenic and Biogenic Organic Aerosols

    SciTech Connect

    Madronich, Sasha

    2015-12-09

    The atmospheric burden of Secondary Organic Aerosols (SOA) remains one of the most important yet uncertain aspects of the radiative forcing of climate. This grant focused on improving our quantitative understanding of SOA formation and evolution, by developing, applying, and improving a highly detailed model of atmospheric organic chemistry, the Generation of Explicit Chemistry and Kinetics of Organics in the Atmosphere (GECKO-A) model. Eleven (11) publications have resulted from this grant.

  13. A Simulation Model for Organic Evolution.

    ERIC Educational Resources Information Center

    Sepe, Stephen

    1988-01-01

    Reports on a process which allows active study of mutating the characteristics of a population of organisms. States that the program, written in C, should run on a MS/DOS system with 640K; CGA is required for graphics output. Describes implementation procedures and two examples using the program. (RT)

  14. Populating a Control Point Database: A cooperative effort between the USGS, Grand Canyon Monitoring and Research Center and the Grand Canyon Youth Organization

    NASA Astrophysics Data System (ADS)

    Brown, K. M.; Fritzinger, C.; Wharton, E.

    2004-12-01

    The Grand Canyon Monitoring and Research Center measures the effects of Glen Canyon Dam operations on the resources along the Colorado River from Glen Canyon Dam to Lake Mead in support of the Grand Canyon Adaptive Management Program. Control points are integral for geo-referencing the myriad of data collected in the Grand Canyon including aerial photography, topographic and bathymetric data used for classification and change-detection analysis of physical, biologic and cultural resources. The survey department has compiled a list of 870 control points installed by various organizations needing to establish a consistent reference for data collected at field sites along the 240 mile stretch of Colorado River in the Grand Canyon. This list is the foundation for the Control Point Database established primarily for researchers, to locate control points and independently geo-reference collected field data. The database has the potential to be a valuable mapping tool for assisting researchers to easily locate a control point and reduce the occurrance of unknowingly installing new control points within close proximity of an existing control point. The database is missing photographs and accurate site description information. Current site descriptions do not accurately define the location of the point but refer to the project that used the point, or some other interesting fact associated with the point. The Grand Canyon Monitoring and Research Center (GCMRC) resolved this problem by turning the data collection effort into an educational exercise for the participants of the Grand Canyon Youth organization. Grand Canyon Youth is a non-profit organization providing experiential education for middle and high school aged youth. GCMRC and the Grand Canyon Youth formed a partnership where GCMRC provided the logistical support, equipment, and training to conduct the field work, and the Grand Canyon Youth provided the time and personnel to complete the field work. Two data

  15. The MEXICO project (Model Experiments in Controlled Conditions): The database and first results of data processing and interpretation

    NASA Astrophysics Data System (ADS)

    Snel, H.; Schepers, J. G.; Montgomerie, B.

    2007-07-01

    The Mexico (Model experiments in Controlled Conditions) was a FP5 project, partly financed by European Commission. The main objective was to create a database of detailed aerodynamic and load measurements on a wind turbine model, in a large and high quality wind tunnel, to be used for model validation and improvement. Here model stands for both the extended BEM modelling used in state-of-the-art design and certification software, and CFD modelling of the rotor and near wake flow. For this purpose a three bladed 4.5 m diameter wind tunnel model was built and instrumented. The wind tunnel experiments were carried out in the open section (9.5*9.5 m2) of the Large Scale Facility of the DNW (German-Netherlands) during a six day campaign in December 2006. The conditions for measurements cover three operational tip speed ratios, many blade pitch angles, three yaw misalignment angles and a small number of unsteady cases in the form of pitch ramps and rotor speed ramps. One of the most important feats of the measurement program was the flow field mapping, with stereo PIV techniques. Overall the measurement campaign was very successful. The paper describes the now existing database and discusses a number of highlights from early data processing and interpretation. It should be stressed that all results are first results, no tunnel correction has been performed so far, nor has the necessary checking of data quality.

  16. PROCARB: A Database of Known and Modelled Carbohydrate-Binding Protein Structures with Sequence-Based Prediction Tools

    PubMed Central

    Malik, Adeel; Firoz, Ahmad; Jha, Vivekanand; Ahmad, Shandar

    2010-01-01

    Understanding of the three-dimensional structures of proteins that interact with carbohydrates covalently (glycoproteins) as well as noncovalently (protein-carbohydrate complexes) is essential to many biological processes and plays a significant role in normal and disease-associated functions. It is important to have a central repository of knowledge available about these protein-carbohydrate complexes as well as preprocessed data of predicted structures. This can be significantly enhanced by tools de novo which can predict carbohydrate-binding sites for proteins in the absence of structure of experimentally known binding site. PROCARB is an open-access database comprising three independently working components, namely, (i) Core PROCARB module, consisting of three-dimensional structures of protein-carbohydrate complexes taken from Protein Data Bank (PDB), (ii) Homology Models module, consisting of manually developed three-dimensional models of N-linked and O-linked glycoproteins of unknown three-dimensional structure, and (iii) CBS-Pred prediction module, consisting of web servers to predict carbohydrate-binding sites using single sequence or server-generated PSSM. Several precomputed structural and functional properties of complexes are also included in the database for quick analysis. In particular, information about function, secondary structure, solvent accessibility, hydrogen bonds and literature reference, and so forth, is included. In addition, each protein in the database is mapped to Uniprot, Pfam, PDB, and so forth. PMID:20671979

  17. Public Opinion Poll Question Databases: An Evaluation

    ERIC Educational Resources Information Center

    Woods, Stephen

    2007-01-01

    This paper evaluates five polling resource: iPOLL, Polling the Nations, Gallup Brain, Public Opinion Poll Question Database, and Polls and Surveys. Content was evaluated on disclosure standards from major polling organizations, scope on a model for public opinion polls, and presentation on a flow chart discussing search limitations and usability.

  18. Exploring Organic Mechanistic Puzzles with Molecular Modeling

    ERIC Educational Resources Information Center

    Horowitz, Gail; Schwartz, Gary

    2004-01-01

    The molecular modeling was used to reinforce more general skills such as deducing and drawing reaction mechanisms, analyzing reaction kinetics and thermodynamics and drawing reaction coordinate energy diagrams. This modeling was done through the design of mechanistic puzzles, involving reactions not familiar to the students.

  19. Using infrared thermography for the creation of a window surface temperature database to validate computer heat transfer models

    SciTech Connect

    Beck, F.A.; Griffith, B.T.; Tuerler, D.; Arasteh, D.

    1995-04-01

    IR thermography is well suited for resolving small differences in the thermal performance of highly insulating window systems. Infrared thermographic measurements made in conjunction with reference emitter techniques in a controlled and characterized laboratory setting can have an absolute accuracy of {plus_minus}0.5{degree}C. Quantitative infrared thermography requires that a number of sources of error related to measurement accuracy and test environmental conditions be quantified and minimized to the extent possible. Laboratory-based infrared thermography can be used to generate window surface temperature profile databases which can be used to direct the development of 2-D and 3-D finite element and finite difference method fenestration heat transfer simulation codes, identify their strengths and weaknesses, set research priorities, and validate finished modeling tools. Development of such a database is under way at Lawrence Berkeley Laboratory, and will be made available for public use.

  20. The GED4GEM project: development of a Global Exposure Database for the Global Earthquake Model initiative

    USGS Publications Warehouse

    Gamba, P.; Cavalca, D.; Jaiswal, K.S.; Huyck, C.; Crowley, H.

    2012-01-01

    In order to quantify earthquake risk of any selected region or a country of the world within the Global Earthquake Model (GEM) framework (www.globalquakemodel.org/), a systematic compilation of building inventory and population exposure is indispensable. Through the consortium of leading institutions and by engaging the domain-experts from multiple countries, the GED4GEM project has been working towards the development of a first comprehensive publicly available Global Exposure Database (GED). This geospatial exposure database will eventually facilitate global earthquake risk and loss estimation through GEM’s OpenQuake platform. This paper provides an overview of the GED concepts, aims, datasets, and inference methodology, as well as the current implementation scheme, status and way forward.

  1. Nearly data-based optimal control for linear discrete model-free systems with delays via reinforcement learning

    NASA Astrophysics Data System (ADS)

    Zhang, Jilie; Zhang, Huaguang; Wang, Binrui; Cai, Tiaoyang

    2016-05-01

    In this paper, a nearly data-based optimal control scheme is proposed for linear discrete model-free systems with delays. The nearly optimal control can be obtained using only measured input/output data from systems, by reinforcement learning technology, which combines Q-learning with value iterative algorithm. First, we construct a state estimator by using the measured input/output data. Second, the quadratic functional is used to approximate the value function at each point in the state space, and the data-based control is designed by Q-learning method using the obtained state estimator. Then, the paper states the method, that is, how to solve the optimal inner kernel matrix ? in the least-square sense, by value iteration algorithm. Finally, the numerical examples are given to illustrate the effectiveness of our approach.

  2. UGTA Photograph Database

    SciTech Connect

    NSTec Environmental Restoration

    2009-04-20

    One of the advantages of the Nevada Test Site (NTS) is that most of the geologic and hydrologic features such as hydrogeologic units (HGUs), hydrostratigraphic units (HSUs), and faults, which are important aspects of flow and transport modeling, are exposed at the surface somewhere in the vicinity of the NTS and thus are available for direct observation. However, due to access restrictions and the remote locations of many of the features, most Underground Test Area (UGTA) participants cannot observe these features directly in the field. Fortunately, National Security Technologies, LLC, geologists and their predecessors have photographed many of these features through the years. During fiscal year 2009, work was done to develop an online photograph database for use by the UGTA community. Photographs were organized, compiled, and imported into Adobe® Photoshop® Elements 7. The photographs were then assigned keyword tags such as alteration type, HGU, HSU, location, rock feature, rock type, and stratigraphic unit. Some fully tagged photographs were then selected and uploaded to the UGTA website. This online photograph database provides easy access for all UGTA participants and can help “ground truth” their analytical and modeling tasks. It also provides new participants a resource to more quickly learn the geology and hydrogeology of the NTS.

  3. Modeling the influence of organic acids on soil weathering

    USGS Publications Warehouse

    Lawrence, Corey R.; Harden, Jennifer W.; Maher, Kate

    2014-01-01

    Biological inputs and organic matter cycling have long been regarded as important factors in the physical and chemical development of soils. In particular, the extent to which low molecular weight organic acids, such as oxalate, influence geochemical reactions has been widely studied. Although the effects of organic acids are diverse, there is strong evidence that organic acids accelerate the dissolution of some minerals. However, the influence of organic acids at the field-scale and over the timescales of soil development has not been evaluated in detail. In this study, a reactive-transport model of soil chemical weathering and pedogenic development was used to quantify the extent to which organic acid cycling controls mineral dissolution rates and long-term patterns of chemical weathering. Specifically, oxalic acid was added to simulations of soil development to investigate a well-studied chronosequence of soils near Santa Cruz, CA. The model formulation includes organic acid input, transport, decomposition, organic-metal aqueous complexation and mineral surface complexation in various combinations. Results suggest that although organic acid reactions accelerate mineral dissolution rates near the soil surface, the net response is an overall decrease in chemical weathering. Model results demonstrate the importance of organic acid input concentrations, fluid flow, decomposition and secondary mineral precipitation rates on the evolution of mineral weathering fronts. In particular, model soil profile evolution is sensitive to kaolinite precipitation and oxalate decomposition rates. The soil profile-scale modeling presented here provides insights into the influence of organic carbon cycling on soil weathering and pedogenesis and supports the need for further field-scale measurements of the flux and speciation of reactive organic compounds.

  4. Modeling the influence of organic acids on soil weathering

    NASA Astrophysics Data System (ADS)

    Lawrence, Corey; Harden, Jennifer; Maher, Kate

    2014-08-01

    Biological inputs and organic matter cycling have long been regarded as important factors in the physical and chemical development of soils. In particular, the extent to which low molecular weight organic acids, such as oxalate, influence geochemical reactions has been widely studied. Although the effects of organic acids are diverse, there is strong evidence that organic acids accelerate the dissolution of some minerals. However, the influence of organic acids at the field-scale and over the timescales of soil development has not been evaluated in detail. In this study, a reactive-transport model of soil chemical weathering and pedogenic development was used to quantify the extent to which organic acid cycling controls mineral dissolution rates and long-term patterns of chemical weathering. Specifically, oxalic acid was added to simulations of soil development to investigate a well-studied chronosequence of soils near Santa Cruz, CA. The model formulation includes organic acid input, transport, decomposition, organic-metal aqueous complexation and mineral surface complexation in various combinations. Results suggest that although organic acid reactions accelerate mineral dissolution rates near the soil surface, the net response is an overall decrease in chemical weathering. Model results demonstrate the importance of organic acid input concentrations, fluid flow, decomposition and secondary mineral precipitation rates on the evolution of mineral weathering fronts. In particular, model soil profile evolution is sensitive to kaolinite precipitation and oxalate decomposition rates. The soil profile-scale modeling presented here provides insights into the influence of organic carbon cycling on soil weathering and pedogenesis and supports the need for further field-scale measurements of the flux and speciation of reactive organic compounds.

  5. NATIONAL URBAN DATABASE AND ACCESS PORTAL TOOL (NUDAPT): FACILITATING ADVANCEMENTS IN URBAN METEOROLOGY AND CLIMATE MODELING WITH COMMUNITY-BASED URBAN DATABASES

    EPA Science Inventory

    We discuss the initial design and application of the National Urban Database and Access Portal Tool (NUDAPT). This new project is sponsored by the USEPA and involves collaborations and contributions from many groups from federal and state agencies, and from private and academic i...

  6. Immediate Dissemination of Student Discoveries to a Model Organism Database Enhances Classroom-Based Research Experiences

    ERIC Educational Resources Information Center

    Wiley, Emily A.; Stover, Nicholas A.

    2014-01-01

    Use of inquiry-based research modules in the classroom has soared over recent years, largely in response to national calls for teaching that provides experience with scientific processes and methodologies. To increase the visibility of in-class studies among interested researchers and to strengthen their impact on student learning, we have…

  7. Collaborative multi organ segmentation by integrating deformable and graphical models.

    PubMed

    Uzunbaş, Mustafa Gökhan; Chen, Chao; Zhang, Shaoting; Poh, Kilian M; Li, Kang; Metaxas, Dimitris

    2013-01-01

    Organ segmentation is a challenging problem on which significant progress has been made. Deformable models (DM) and graphical models (GM) are two important categories of optimization based image segmentation methods. Efforts have been made on integrating two types of models into one framework. However, previous methods are not designed for segmenting multiple organs simultaneously and accurately. In this paper, we propose a hybrid multi organ segmentation approach by integrating DM and GM in a coupled optimization framework. Specifically, we show that region-based deformable models can be integrated with Markov Random Fields (MRF), such that multiple models' evolutions are driven by a maximum a posteriori (MAP) inference. It brings global and local deformation constraints into a unified framework for simultaneous segmentation of multiple objects in an image. We validate this proposed method on two challenging problems of multi organ segmentation, and the results are promising. PMID:24579136

  8. Databases for Microbiologists

    DOE PAGESBeta

    Zhulin, Igor B.

    2015-05-26

    Databases play an increasingly important role in biology. They archive, store, maintain, and share information on genes, genomes, expression data, protein sequences and structures, metabolites and reactions, interactions, and pathways. All these data are critically important to microbiologists. Furthermore, microbiology has its own databases that deal with model microorganisms, microbial diversity, physiology, and pathogenesis. Thousands of biological databases are currently available, and it becomes increasingly difficult to keep up with their development. Finally, the purpose of this minireview is to provide a brief survey of current databases that are of interest to microbiologists.

  9. Databases for Microbiologists

    PubMed Central

    2015-01-01

    Databases play an increasingly important role in biology. They archive, store, maintain, and share information on genes, genomes, expression data, protein sequences and structures, metabolites and reactions, interactions, and pathways. All these data are critically important to microbiologists. Furthermore, microbiology has its own databases that deal with model microorganisms, microbial diversity, physiology, and pathogenesis. Thousands of biological databases are currently available, and it becomes increasingly difficult to keep up with their development. The purpose of this minireview is to provide a brief survey of current databases that are of interest to microbiologists. PMID:26013493

  10. Risk Factors for the Development of Heterotopic Ossification in Seriously Burned Adults: A NIDRR Burn Model System Database Analysis

    PubMed Central

    Levi, Benjamin; Jayakumar, Prakash; Giladi, Avi; Jupiter, Jesse B.; Ring, David C.; Kowalske, Karen; Gibran, Nicole S.; Herndon, David; Schneider, Jeffrey C.; Ryan, Colleen M.

    2015-01-01

    Purpose Heterotopic ossification (HO) is a debilitating complication of burn injury; however, incidence and risk factors are poorly understood. In this study we utilize a multicenter database of adults with burn injuries to identify and analyze clinical factors that predict HO formation. Methods Data from 6 high-volume burn centers, in the Burn Injury Model System Database, were analyzed. Univariate logistic regression models were used for model selection. Cluster-adjusted multivariate logistic regression was then used to evaluate the relationship between clinical and demographic data and the development of HO. Results Of 2,979 patients in the database with information on HO that addressed risk factors for development of HO, 98 (3.5%) developed HO. Of these 98 patients, 97 had arm burns, and 96 had arm grafts. Controlling for age and sex in a multivariate model, patients with >30% total body surface area (TBSA) burn had 11.5x higher odds of developing HO (p<0.001), and those with arm burns that required skin grafting had 96.4x higher odds of developing HO (p=0.04). For each additional time a patient went to the operating room, odds of HO increased 30% (OR 1.32, p<0.001), and each additional ventilator day increase odds 3.5% (OR 1.035, p<0.001). Joint contracture, inhalation injury, and bone exposure did not significantly increase odds of HO. Conclusion Risk factors for HO development include >30% TBSA burn, arm burns, arm grafts, ventilator days, and number of trips to the operating room. Future studies can use these results to identify highest-risk patients to guide deployment of prophylactic and experimental treatments. PMID:26496115

  11. Elimination kinetic model for organic chemicals in earthworms.

    PubMed

    Dimitrova, N; Dimitrov, S; Georgieva, D; Van Gestel, C A M; Hankard, P; Spurgeon, D; Li, H; Mekenyan, O

    2010-08-15

    Mechanistic understanding of bioaccumulation in different organisms and environments should take into account the influence of organism and chemical depending factors on the uptake and elimination kinetics of chemicals. Lipophilicity, metabolism, sorption (bioavailability) and biodegradation of chemicals are among the important factors that may significantly affect the bioaccumulation process in soil organisms. This study attempts to model elimination kinetics of organic chemicals in earthworms by accounting for the effects of both chemical and biological properties, including metabolism. The modeling approach that has been developed is based on the concept for simulating metabolism used in the BCF base-line model developed for predicting bioaccumulation in fish. Metabolism was explicitly accounted for by making use of the TIMES engine for simulation of metabolism and a set of principal transformations. Kinetic characteristics of transformations were estimated on the basis of observed kinetics data for the elimination of organic chemicals from earthworms. PMID:20185163

  12. Nematodes: Model Organisms in High School Biology

    ERIC Educational Resources Information Center

    Bliss, TJ; Anderson, Margery; Dillman, Adler; Yourick, Debra; Jett, Marti; Adams, Byron J.; Russell, RevaBeth

    2007-01-01

    In a collaborative effort between university researchers and high school science teachers, an inquiry-based laboratory module was designed using two species of insecticidal nematodes to help students apply scientific inquiry and elements of thoughtful experimental design. The learning experience and model are described in this article. (Contains 4…

  13. HITEMP derived spectral database for the prediction of jet engine exhaust infrared emission using a statistical band model

    NASA Astrophysics Data System (ADS)

    Lindermeir, E.; Beier, K.

    2012-08-01

    The spectroscopic database HITEMP 2010 is used to upgrade the parameters of the statistical molecular band model which is part of the infrared signature prediction code NIRATAM (NATO InfraRed Air TArget Model). This band model was recommended by NASA and is applied in several codes that determine the infrared emission of combustion gases. The upgrade regards spectral absorption coefficients and line densities of the gases H2O, CO2, and CO in the spectral region 400-5000 cm-1 (2-25μm) with a spectral resolution of 5 cm-1. The temperature range 100-3000 K is covered. Two methods to update the database are presented: the usually applied method as provided in the literature and an alternative, more laborious procedure that employs least squares fitting. The achieved improvements resulting from both methods are demonstrated by comparisons of radiance spectra obtained from the band model to line-by-line results. The performance in a realistic scenario is investigated on the basis of measured and predicted spectra of a jet aircraft plume in afterburner mode.

  14. Prospective validation of a comprehensive in silico hERG model and its applications to commercial compound and drug databases.

    PubMed

    Doddareddy, Munikumar R; Klaasse, Elisabeth C; Shagufta; Ijzerman, Adriaan P; Bender, Andreas

    2010-05-01

    Ligand-based in silico hERG models were generated for 2 644 compounds using linear discriminant analysis (LDA) and support vector machines (SVM). As a result, the dataset used for the model generation is the largest publicly available (see Supporting Information). Extended connectivity fingerprints (ECFPs) and functional class fingerprints (FCFPs) were used to describe chemical space. All models showed area under curve (AUC) values ranging from 0.89 to 0.94 in a fivefold cross-validation, indicating high model consistency. Models correctly predicted 80 % of an additional, external test set; Y-scrambling was also performed to rule out chance correlation. Additionally models based on patch clamp data and radioligand binding data were generated separately to analyze their predictive ability when compared to combined models. To experimentally validate the models, 50 of the predicted hERG blockers from the Chembridge database and ten of the predicted non-hERG blockers from an in-house compound library were selected for biological evaluation. Out of those 50 predicted hERG blockers, tested at a concentration of 10 microM, 18 compounds showed more than 50 % displacement of [(3)H]astemizole binding to cell membranes expressing the hERG channel. K(i) values of four of the selected binders were determined to be in the micromolar and high nanomolar range (K(i) (VH01)=2.0 microM, K(i) (VH06)=0.15 microM, K(i) (VH19)=1.1 microM and K(i) (VH47)=18 microM). Of these four compounds, VH01 and VH47 showed also a second, even higher affinity binding site with K(i) values of 7.4 nM and 36 nM, respectively. In the case of non-hERG blockers, all ten compounds tested were found to be inactive, showing less than 50 % displacement of [(3)H]astemizole binding at 10 microM. These experimentally validated models were then used to virtually screen commercial compound databases to evaluate whether they contain hERG blockers. 109 784 (23 %) of Chembridge, 133 175 (38 %) of Chemdiv, 111 737 (31

  15. VibrioBase: a model for next-generation genome and annotation database development.

    PubMed

    Choo, Siew Woh; Heydari, Hamed; Tan, Tze King; Siow, Cheuk Chuen; Beh, Ching Yew; Wee, Wei Yee; Mutha, Naresh V R; Wong, Guat Jah; Ang, Mia Yang; Yazdi, Amir Hessam

    2014-01-01

    To facilitate the ongoing research of Vibrio spp., a dedicated platform for the Vibrio research community is needed to host the fast-growing amount of genomic data and facilitate the analysis of these data. We present VibrioBase, a useful resource platform, providing all basic features of a sequence database with the addition of unique analysis tools which could be valuable for the Vibrio research community. VibrioBase currently houses a total of 252 Vibrio genomes developed in a user-friendly manner and useful to enable the analysis of these genomic data, particularly in the field of comparative genomics. Besides general data browsing features, VibrioBase offers analysis tools such as BLAST interfaces and JBrowse genome browser. Other important features of this platform include our newly developed in-house tools, the pairwise genome comparison (PGC) tool, and pathogenomics profiling tool (PathoProT). The PGC tool is useful in the identification and comparative analysis of two genomes, whereas PathoProT is designed for comparative pathogenomics analysis of Vibrio strains. Both of these tools will enable researchers with little experience in bioinformatics to get meaningful information from Vibrio genomes with ease. We have tested the validity and suitability of these tools and features for use in the next-generation database development.

  16. Machine learning for toxicity characterization of organic chemical emissions using USEtox database: Learning the structure of the input space.

    PubMed

    Marvuglia, Antonino; Kanevski, Mikhail; Benetto, Enrico

    2015-10-01

    Toxicity characterization of chemical emissions in Life Cycle Assessment (LCA) is a complex task which usually proceeds via multimedia (fate, exposure and effect) models attached to models of dose-response relationships to assess the effects on target. Different models and approaches do exist, but all require a vast amount of data on the properties of the chemical compounds being assessed, which are hard to collect or hardly publicly available (especially for thousands of less common or newly developed chemicals), therefore hampering in practice the assessment in LCA. An example is USEtox, a consensual model for the characterization of human toxicity and freshwater ecotoxicity. This paper places itself in a line of research aiming at providing a methodology to reduce the number of input parameters necessary to run multimedia fate models, focusing in particular to the application of the USEtox toxicity model. By focusing on USEtox, in this paper two main goals are pursued: 1) performing an extensive exploratory analysis (using dimensionality reduction techniques) of the input space constituted by the substance-specific properties at the aim of detecting particular patterns in the data manifold and estimating the dimension of the subspace in which the data manifold actually lies; and 2) exploring the application of a set of linear models, based on partial least squares (PLS) regression, as well as a nonlinear model (general regression neural network--GRNN) in the seek for an automatic selection strategy of the most informative variables according to the modelled output (USEtox factor). After extensive analysis, the intrinsic dimension of the input manifold has been identified between three and four. The variables selected as most informative may vary according to the output modelled and the model used, but for the toxicity factors modelled in this paper the input variables selected as most informative are coherent with prior expectations based on scientific knowledge

  17. Machine learning for toxicity characterization of organic chemical emissions using USEtox database: Learning the structure of the input space.

    PubMed

    Marvuglia, Antonino; Kanevski, Mikhail; Benetto, Enrico

    2015-10-01

    Toxicity characterization of chemical emissions in Life Cycle Assessment (LCA) is a complex task which usually proceeds via multimedia (fate, exposure and effect) models attached to models of dose-response relationships to assess the effects on target. Different models and approaches do exist, but all require a vast amount of data on the properties of the chemical compounds being assessed, which are hard to collect or hardly publicly available (especially for thousands of less common or newly developed chemicals), therefore hampering in practice the assessment in LCA. An example is USEtox, a consensual model for the characterization of human toxicity and freshwater ecotoxicity. This paper places itself in a line of research aiming at providing a methodology to reduce the number of input parameters necessary to run multimedia fate models, focusing in particular to the application of the USEtox toxicity model. By focusing on USEtox, in this paper two main goals are pursued: 1) performing an extensive exploratory analysis (using dimensionality reduction techniques) of the input space constituted by the substance-specific properties at the aim of detecting particular patterns in the data manifold and estimating the dimension of the subspace in which the data manifold actually lies; and 2) exploring the application of a set of linear models, based on partial least squares (PLS) regression, as well as a nonlinear model (general regression neural network--GRNN) in the seek for an automatic selection strategy of the most informative variables according to the modelled output (USEtox factor). After extensive analysis, the intrinsic dimension of the input manifold has been identified between three and four. The variables selected as most informative may vary according to the output modelled and the model used, but for the toxicity factors modelled in this paper the input variables selected as most informative are coherent with prior expectations based on scientific knowledge

  18. Representational Translation with Concrete Models in Organic Chemistry

    ERIC Educational Resources Information Center

    Stull, Andrew T.; Hegarty, Mary; Dixon, Bonnie; Stieff, Mike

    2012-01-01

    In representation-rich domains such as organic chemistry, students must be facile and accurate when translating between different 2D representations, such as diagrams. We hypothesized that translating between organic chemistry diagrams would be more accurate when concrete models were used because difficult mental processes could be augmented by…

  19. Expatriate Training in International Nongovernmental Organizations: A Model for Research

    ERIC Educational Resources Information Center

    Chang, Wei-Wen

    2005-01-01

    In light of the massive tsunami relief efforts that were still being carried out by humanitarian organizations around the world when this article went to press, this article points out a lack of human resources development research in international nongovernmental organizations (INGOs) and proposes a conceptual model for future empirical research.…

  20. Resilient organizations: matrix model and service line management.

    PubMed

    Westphal, Judith A

    2005-09-01

    Resilient organizations modify structures to meet the demands of the marketplace. The author describes a structure that enables multihospital organizations to innovate and rapidly adapt to changes. Service line management within a matrix model is an evolving organizational structure for complex systems in which nurses are pivotal members.

  1. Self-organizing map models of language acquisition

    PubMed Central

    Li, Ping; Zhao, Xiaowei

    2013-01-01

    Connectionist models have had a profound impact on theories of language. While most early models were inspired by the classic parallel distributed processing architecture, recent models of language have explored various other types of models, including self-organizing models for language acquisition. In this paper, we aim at providing a review of the latter type of models, and highlight a number of simulation experiments that we have conducted based on these models. We show that self-organizing connectionist models can provide significant insights into long-standing debates in both monolingual and bilingual language development. We suggest future directions in which these models can be extended, to better connect with behavioral and neural data, and to make clear predictions in testing relevant psycholinguistic theories. PMID:24312061

  2. Self-organizing map models of language acquisition.

    PubMed

    Li, Ping; Zhao, Xiaowei

    2013-01-01

    Connectionist models have had a profound impact on theories of language. While most early models were inspired by the classic parallel distributed processing architecture, recent models of language have explored various other types of models, including self-organizing models for language acquisition. In this paper, we aim at providing a review of the latter type of models, and highlight a number of simulation experiments that we have conducted based on these models. We show that self-organizing connectionist models can provide significant insights into long-standing debates in both monolingual and bilingual language development. We suggest future directions in which these models can be extended, to better connect with behavioral and neural data, and to make clear predictions in testing relevant psycholinguistic theories.

  3. Phase Equilibria Diagrams Database

    National Institute of Standards and Technology Data Gateway

    SRD 31 NIST/ACerS Phase Equilibria Diagrams Database (PC database for purchase)   The Phase Equilibria Diagrams Database contains commentaries and more than 21,000 diagrams for non-organic systems, including those published in all 21 hard-copy volumes produced as part of the ACerS-NIST Phase Equilibria Diagrams Program (formerly titled Phase Diagrams for Ceramists): Volumes I through XIV (blue books); Annuals 91, 92, 93; High Tc Superconductors I & II; Zirconium & Zirconia Systems; and Electronic Ceramics I. Materials covered include oxides as well as non-oxide systems such as chalcogenides and pnictides, phosphates, salt systems, and mixed systems of these classes.

  4. A REVIEW OF BIOACCUMULATION MODELING APPROACHES FOR PERSISTENT ORGANIC POLLUTANTS

    EPA Science Inventory

    Persistent organic pollutants and mercury are likely to bioaccumulate in biological components of the environment, including fish and wildlife. The complex and long-term dynamics involved with bioaccumulation are often represented with models. Current scientific developments in t...

  5. Ecosystem models. (Latest citations from the NTIS Bibliographic database). Published Search

    SciTech Connect

    Not Available

    1993-11-01

    The bibliography contains citations concerning the design and applications of ecosystem models. Ecosystem simulation and characterization models, together with models for marine biology, plants, microorganisms, and food chains, are described. Models that assess the effect of pollutants on specific environments and habitat suitability index models are also included. (Contains 250 citations and includes a subject term index and title list.)

  6. Knowledge-based model of hydrogen-bonding propensity in organic crystals.

    PubMed

    Galek, Peter T A; Fábián, László; Motherwell, W D Samuel; Allen, Frank H; Feeder, Neil

    2007-10-01

    A new method is presented to predict which donors and acceptors form hydrogen bonds in a crystal structure, based on the statistical analysis of hydrogen bonds in the Cambridge Structural Database (CSD). The method is named the logit hydrogen-bonding propensity (LHP) model. The approach has a potential application in identifying both likely and unusual hydrogen bonding, which can help to rationalize stable and metastable crystalline forms, of relevance to drug development in the pharmaceutical industry. Whilst polymorph prediction techniques are widely used, the LHP model is knowledge-based and is not restricted by the computational issues of polymorph prediction, and as such may form a valuable precursor to polymorph screening. Model construction applies logistic regression, using training data obtained with a new survey method based on the CSD system. The survey categorizes the hydrogen bonds and extracts model parameter values using descriptive structural and chemical properties from three-dimensional organic crystal structures. LHP predictions from a fitted model are made using two-dimensional observables alone. In the initial cases analysed, the model is highly accurate, achieving approximately 90% correct classification of both observed hydrogen bonds and non-interacting donor-acceptor pairs. Extensive statistical validation shows the LHP model to be robust across a range of small-molecule organic crystal structures. PMID:17873446

  7. Maize databases

    Technology Transfer Automated Retrieval System (TEKTRAN)

    This chapter is a succinct overview of maize data held in the species-specific database MaizeGDB (the Maize Genomics and Genetics Database), and selected multi-species data repositories, such as Gramene/Ensembl Plants, Phytozome, UniProt and the National Center for Biotechnology Information (NCBI), ...

  8. Influence of dissolved organic carbon content on modelling natural organic matter acid-base properties.

    PubMed

    Garnier, Cédric; Mounier, Stéphane; Benaïm, Jean Yves

    2004-10-01

    Natural organic matter (NOM) behaviour towards proton is an important parameter to understand NOM fate in the environment. Moreover, it is necessary to determine NOM acid-base properties before investigating trace metals complexation by natural organic matter. This work focuses on the possibility to determine these acid-base properties by accurate and simple titrations, even at low organic matter concentrations. So, the experiments were conducted on concentrated and diluted solutions of extracted humic and fulvic acid from Laurentian River, on concentrated and diluted model solutions of well-known simple molecules (acetic and phenolic acids), and on natural samples from the Seine river (France) which are not pre-concentrated. Titration experiments were modelled by a 6 acidic-sites discrete model, except for the model solutions. The modelling software used, called PROSECE (Programme d'Optimisation et de SpEciation Chimique dans l'Environnement), has been developed in our laboratory, is based on the mass balance equilibrium resolution. The results obtained on extracted organic matter and model solutions point out a threshold value for a confident determination of the studied organic matter acid-base properties. They also show an aberrant decreasing carboxylic/phenolic ratio with increasing sample dilution. This shift is neither due to any conformational effect, since it is also observed on model solutions, nor to ionic strength variations which is controlled during all experiments. On the other hand, it could be the result of an electrode troubleshooting occurring at basic pH values, which effect is amplified at low total concentration of acidic sites. So, in our conditions, the limit for a correct modelling of NOM acid-base properties is defined as 0.04 meq of total analysed acidic sites concentration. As for the analysed natural samples, due to their high acidic sites content, it is possible to model their behaviour despite the low organic carbon concentration.

  9. Dietary Uptake Models Used for Modeling the Bioaccumulation of Organic Contaminants in Fish

    EPA Science Inventory

    Numerous models have been developed to predict the bioaccumulation of organic chemicals in fish. Although chemical dietary uptake can be modeled using assimilation efficiencies, bioaccumulation models fall into two distinct groups. The first group implicitly assumes that assimila...

  10. Causal biological network database: a comprehensive platform of causal biological network models focused on the pulmonary and vascular systems

    PubMed Central

    Boué, Stéphanie; Talikka, Marja; Westra, Jurjen Willem; Hayes, William; Di Fabio, Anselmo; Park, Jennifer; Schlage, Walter K.; Sewer, Alain; Fields, Brett; Ansari, Sam; Martin, Florian; Veljkovic, Emilija; Kenney, Renee; Peitsch, Manuel C.; Hoeng, Julia

    2015-01-01

    With the wealth of publications and data available, powerful and transparent computational approaches are required to represent measured data and scientific knowledge in a computable and searchable format. We developed a set of biological network models, scripted in the Biological Expression Language, that reflect causal signaling pathways across a wide range of biological processes, including cell fate, cell stress, cell proliferation, inflammation, tissue repair and angiogenesis in the pulmonary and cardiovascular context. This comprehensive collection of networks is now freely available to the scientific community in a centralized web-based repository, the Causal Biological Network database, which is composed of over 120 manually curated and well annotated biological network models and can be accessed at http://causalbionet.com. The website accesses a MongoDB, which stores all versions of the networks as JSON objects and allows users to search for genes, proteins, biological processes, small molecules and keywords in the network descriptions to retrieve biological networks of interest. The content of the networks can be visualized and browsed. Nodes and edges can be filtered and all supporting evidence for the edges can be browsed and is linked to the original articles in PubMed. Moreover, networks may be downloaded for further visualization and evaluation. Database URL: http://causalbionet.com PMID:25887162

  11. Causal biological network database: a comprehensive platform of causal biological network models focused on the pulmonary and vascular systems.

    PubMed

    Boué, Stéphanie; Talikka, Marja; Westra, Jurjen Willem; Hayes, William; Di Fabio, Anselmo; Park, Jennifer; Schlage, Walter K; Sewer, Alain; Fields, Brett; Ansari, Sam; Martin, Florian; Veljkovic, Emilija; Kenney, Renee; Peitsch, Manuel C; Hoeng, Julia

    2015-01-01

    With the wealth of publications and data available, powerful and transparent computational approaches are required to represent measured data and scientific knowledge in a computable and searchable format. We developed a set of biological network models, scripted in the Biological Expression Language, that reflect causal signaling pathways across a wide range of biological processes, including cell fate, cell stress, cell proliferation, inflammation, tissue repair and angiogenesis in the pulmonary and cardiovascular context. This comprehensive collection of networks is now freely available to the scientific community in a centralized web-based repository, the Causal Biological Network database, which is composed of over 120 manually curated and well annotated biological network models and can be accessed at http://causalbionet.com. The website accesses a MongoDB, which stores all versions of the networks as JSON objects and allows users to search for genes, proteins, biological processes, small molecules and keywords in the network descriptions to retrieve biological networks of interest. The content of the networks can be visualized and browsed. Nodes and edges can be filtered and all supporting evidence for the edges can be browsed and is linked to the original articles in PubMed. Moreover, networks may be downloaded for further visualization and evaluation. Database URL: http://causalbionet.com

  12. An isolated line-shape model to go beyond the Voigt profile in spectroscopic databases and radiative transfer codes

    NASA Astrophysics Data System (ADS)

    Ngo, N. H.; Lisak, D.; Tran, H.; Hartmann, J.-M.

    2013-11-01

    We demonstrate that a previously proposed model opens the route for the inclusion of refined non-Voigt profiles in spectroscopic databases and atmospheric radiative transfer codes. Indeed, this model fulfills many essential requirements: (i) it takes both velocity changes and the speed dependences of the pressure-broadening and -shifting coefficients into account. (ii) It leads to accurate descriptions of the line shapes of very different molecular systems. Tests made for pure H2, CO2 and O2 and for H2O diluted in N2 show that residuals are down to ≃0.2% of the peak absorption, (except for the untypical system of H2 where a maximum residual of ±3% is reached), thus fulfilling the precision requirements of the most demanding remote sensing experiments. (iii) It is based on a limited set of parameters for each absorption line that have known dependences on pressure and can thus be stored in databases. (iv) Its calculation requires very reasonable computer costs, only a few times higher than that of a usual Voigt profile. Its inclusion in radiative transfer codes will thus induce bearable CPU time increases. (v) It can be extended in order to take line-mixing effects into account, at least within the so-called first-order approximation.

  13. Organism-level models: When mechanisms and statistics fail us

    NASA Astrophysics Data System (ADS)

    Phillips, M. H.; Meyer, J.; Smith, W. P.; Rockhill, J. K.

    2014-03-01

    Purpose: To describe the unique characteristics of models that represent the entire course of radiation therapy at the organism level and to highlight the uses to which such models can be put. Methods: At the level of an organism, traditional model-building runs into severe difficulties. We do not have sufficient knowledge to devise a complete biochemistry-based model. Statistical model-building fails due to the vast number of variables and the inability to control many of them in any meaningful way. Finally, building surrogate models, such as animal-based models, can result in excluding some of the most critical variables. Bayesian probabilistic models (Bayesian networks) provide a useful alternative that have the advantages of being mathematically rigorous, incorporating the knowledge that we do have, and being practical. Results: Bayesian networks representing radiation therapy pathways for prostate cancer and head & neck cancer were used to highlight the important aspects of such models and some techniques of model-building. A more specific model representing the treatment of occult lymph nodes in head & neck cancer were provided as an example of how such a model can inform clinical decisions. A model of the possible role of PET imaging in brain cancer was used to illustrate the means by which clinical trials can be modelled in order to come up with a trial design that will have meaningful outcomes. Conclusions: Probabilistic models are currently the most useful approach to representing the entire therapy outcome process.

  14. Mutant mice: experimental organisms as materialised models in biomedicine.

    PubMed

    Huber, Lara; Keuck, Lara K

    2013-09-01

    Animal models have received particular attention as key examples of material models. In this paper, we argue that the specificities of establishing animal models-acknowledging their status as living beings and as epistemological tools-necessitate a more complex account of animal models as materialised models. This becomes particularly evident in animal-based models of diseases that only occur in humans: in these cases, the representational relation between animal model and human patient needs to be generated and validated. The first part of this paper presents an account of how disease-specific animal models are established by drawing on the example of transgenic mice models for Alzheimer's disease. We will introduce an account of validation that involves a three-fold process including (1) from human being to experimental organism; (2) from experimental organism to animal model; and (3) from animal model to human patient. This process draws upon clinical relevance as much as scientific practices and results in disease-specific, yet incomplete, animal models. The second part of this paper argues that the incompleteness of models can be described in terms of multi-level abstractions. We qualify this notion by pointing to different experimental techniques and targets of modelling, which give rise to a plurality of models for a specific disease.

  15. Improved AIOMFAC model parameterisation of the temperature dependence of activity coefficients for aqueous organic mixtures

    NASA Astrophysics Data System (ADS)

    Ganbavale, G.; Zuend, A.; Marcolli, C.; Peter, T.

    2015-01-01

    This study presents a new, improved parameterisation of the temperature dependence of activity coefficients in the AIOMFAC (Aerosol Inorganic-Organic Mixtures Functional groups Activity Coefficients) model applicable for aqueous as well as water-free organic solutions. For electrolyte-free organic and organic-water mixtures the AIOMFAC model uses a group-contribution approach based on UNIFAC (UNIversal quasi-chemical Functional-group Activity Coefficients). This group-contribution approach explicitly accounts for interactions among organic functional groups and between organic functional groups and water. The previous AIOMFAC version uses a simple parameterisation of the temperature dependence of activity coefficients, aimed to be applicable in the temperature range from ~ 275 to ~ 400 K. With the goal to improve the description of a wide variety of organic compounds found in atmospheric aerosols, we extend the AIOMFAC parameterisation for the functional groups carboxyl, hydroxyl, ketone, aldehyde, ether, ester, alkyl, aromatic carbon-alcohol, and aromatic hydrocarbon to atmospherically relevant low temperatures. To this end we introduce a new parameterisation for the temperature dependence. The improved temperature dependence parameterisation is derived from classical thermodynamic theory by describing effects from changes in molar enthalpy and heat capacity of a multi-component system. Thermodynamic equilibrium data of aqueous organic and water-free organic mixtures from the literature are carefully assessed and complemented with new measurements to establish a comprehensive database, covering a wide temperature range (~ 190 to ~ 440 K) for many of the functional group combinations considered. Different experimental data types and their processing for the estimation of AIOMFAC model parameters are discussed. The new AIOMFAC parameterisation for the temperature dependence of activity coefficients from low to high temperatures shows an overall improvement of 28% in

  16. The power of an ontology-driven developmental toxicity database for data mining and computational modeling

    EPA Science Inventory

    Modeling of developmental toxicology presents a significant challenge to computational toxicology due to endpoint complexity and lack of data coverage. These challenges largely account for the relatively few modeling successes using the structure–activity relationship (SAR) parad...

  17. Soil moisture modelling of a SMOS pixel: interest of using the PERSIANN database over the Valencia Anchor Station

    NASA Astrophysics Data System (ADS)

    Juglea, S.; Kerr, Y.; Mialon, A.; Lopez-Baeza, E.; Braithwaite, D.; Hsu, K.

    2010-08-01

    In the framework of Soil Moisture and Ocean Salinity (SMOS) Calibration/Validation (Cal/Val) activities, this study addresses the use of the PERSIANN-CCS1database in hydrological applications to accurately simulate a whole SMOS pixel by representing the spatial and temporal heterogeneity of the soil moisture fields over a wide area (50×50 km2). The study focuses on the Valencia Anchor Station (VAS) experimental site, in Spain, which is one of the main SMOS Cal/Val sites in Europe. A faithful representation of the soil moisture distribution at SMOS pixel scale (50×50 km2) requires an accurate estimation of the amount and temporal/spatial distribution of precipitation. To quantify the gain of using the comprehensive PERSIANN database instead of sparsely distributed rain gauge measurements, comparisons between in situ observations and satellite rainfall data are done both at point and areal scale. An overestimation of the satellite rainfall amounts is observed in most of the cases (about 66%) but the precipitation occurrences are in general retrieved (about 67%). To simulate the high variability in space and time of surface soil moisture, a Soil Vegetation Atmosphere Transfer (SVAT) model - ISBA (Interactions between Soil Biosphere Atmosphere) is used. The interest of using satellite rainfall estimates as well as the influence that the precipitation events can induce on the modelling of the water content in the soil is depicted by a comparison between different soil moisture data. Point-like and spatialized simulated data using rain gauge observations or PERSIANN - CCS database as well as ground measurements are used. It is shown that a good adequacy is reached in most part of the year, the precipitation differences having less impact upon the simulated soil moisture. The behaviour of simulated surface soil moisture at SMOS scale is verified by the use of remote sensing data from the Advanced Microwave Scanning Radiometer on Earth observing System (AMSR-E). We show

  18. An Online Database and User Community for Physical Models in the Engineering Classroom

    ERIC Educational Resources Information Center

    Welch, Robert W.; Klosky, J. Ledlie

    2007-01-01

    This paper will present information about the Web site--www.handsonmechanics.com, the process to develop the Web site, the vetting and management process for inclusion of physical models by the faculty at West Point, and how faculty at other institutions can add physical models and participate in the site as it grows. Each physical model has a…

  19. An Application of the Programmatic Organization Model to Applied Research.

    ERIC Educational Resources Information Center

    Haring, Norris G.; Krug, David A.

    Presented is a programatic organization model which is applied to research needs in special education. Major advantages of the model are given to be that of providing for several research activities which focus on one common problem, reducing the duplication of research effort, providing greater efficiency in use of research personnel and…

  20. Healing models for organizations: description, measurement, and outcomes.

    PubMed

    Malloch, K

    2000-01-01

    Healthcare leaders are continually searching for ways to improve their ability to provide optimal healthcare services, be financially viable, and retain quality caregivers, often feeling like such goals are impossible to achieve in today's intensely competitive environment. Many healthcare leaders intuitively recognize the need for more humanistic models and the probable connection with positive patient outcomes and financial success but are hesitant to make significant changes in their organizations because of the lack of model descriptions or documented recognition of the clinical and financial advantages of humanistic models. This article describes a study that was developed in response to the increasing work in humanistic or healing environment models and the need for validation of the advantages of such models. The healthy organization model, a framework for healthcare organizations that incorporates humanistic healing values within the traditional structure, is presented as a result of the study. This model addresses the importance of optimal clinical services, financial performance, and staff satisfaction. The five research-based organizational components that form the framework are described, and key indicators of organizational effectiveness over a five-year period are presented. The resulting empirical data are strongly supportive of the healing model and reflect positive outcomes for the organization.

  1. Simple model of self-organized biological evolution

    SciTech Connect

    de Boer, J.; Derrida, B.; Flyvbjerg, H.; Jackson, A.D.; Wettig, T. The Isaac Newton Institute for Mathematical Sciences, 20 Clarkson Road, Cambridge, CB4 0EH Laboratoire de Physique Statistique, Ecole Normale Superieure, 24 rue Lhomond, F-75005 Paris Service de Physique Theorique, Centre de Etudes Nucleaires de Saclay, F-91191, Gif-Sur-Yvette CONNECT, The Niels Bohr Institute, Blegdamsvej 17, DK-2100 Copenhagen )

    1994-08-08

    We give an exact solution of a recently proposed self-organized critical model of biological evolution. We show that the model has a power law distribution of durations of coevolutionary avalanches'' with a mean field exponent 3/2. We also calculate analytically the finite size effects which cut off this power law at times of the order of the system size.

  2. Institutionalizing Innovation in an Organization: A Model and Case Study.

    ERIC Educational Resources Information Center

    Slawski, Carl

    A policy systems theoretical analysis of the problem of institutionalizing innovation in an organization is summarized in a flow diagram. The model is presented in terms of specific hypotheses, and then illustrated with a case of frustrated innovation, the 1968-69 crisis and strike at San Francisco State College. The model is set up (1) to help…

  3. A Workforce Design Model: Providing Energy to Organizations in Transition

    ERIC Educational Resources Information Center

    Halm, Barry J.

    2011-01-01

    The purpose of this qualitative study was to examine the change in performance realized by a professional services organization, which resulted in the Life Giving Workforce Design (LGWD) model through a grounded theory research design. This study produced a workforce design model characterized as an organizational blueprint that provides virtuous…

  4. Principles of chromatin organization in yeast: relevance of polymer models to describe nuclear organization and dynamics.

    PubMed

    Wang, Renjie; Mozziconacci, Julien; Bancaud, Aurélien; Gadal, Olivier

    2015-06-01

    Nuclear organization can impact on all aspects of the genome life cycle. This organization is thoroughly investigated by advanced imaging and chromosome conformation capture techniques, providing considerable amount of datasets describing the spatial organization of chromosomes. In this review, we will focus on polymer models to describe chromosome statics and dynamics in the yeast Saccharomyces cerevisiae. We suggest that the equilibrium configuration of a polymer chain tethered at both ends and placed in a confined volume is consistent with the current literature, implying that local chromatin interactions play a secondary role in yeast nuclear organization. Future challenges are to reach an integrated multi-scale description of yeast chromosome organization, which is crucially needed to improve our understanding of the regulation of genomic transaction. PMID:25956973

  5. Spatial Arrangment of Organic Compounds on a Model Mineral Surface: Implications for Soil Organic Matter Stabilization

    SciTech Connect

    Petridis, Loukas; Ambaye, Haile Arena; Jagadamma, Sindhu; Kilbey, S. Michael; Lokitz, Bradley S; Lauter, Valeria; Mayes, Melanie

    2014-01-01

    The complexity of the mineral organic carbon interface may influence the extent of stabilization of organic carbon compounds in soils, which is important for global climate futures. The nanoscale structure of a model interface was examined here by depositing films of organic carbon compounds of contrasting chemical character, hydrophilic glucose and amphiphilic stearic acid, onto a soil mineral analogue (Al2O3). Neutron reflectometry, a technique which provides depth-sensitive insight into the organization of the thin films, indicates that glucose molecules reside in a layer between Al2O3 and stearic acid, a result that was verified by water contact angle measurements. Molecular dynamics simulations reveal the thermodynamic driving force behind glucose partitioning on the mineral interface: The entropic penalty of confining the less mobile glucose on the mineral surface is lower than for stearic acid. The fundamental information obtained here helps rationalize how complex arrangements of organic carbon on soil mineral surfaces may arise

  6. Self-Organized Criticality in an Asexual Model?

    NASA Astrophysics Data System (ADS)

    Chisholm, Colin; Jan, Naeem; Gibbs, Peter; Erzan, Ayşe.

    Recent work has shown that the distribution of steady state mutations for an asexual ``bacteria'' model has features similar to that seen in Self-Organized Critical (SOC) sandpile model of Bak et al. We investigate this coincidence further and search for ``self-organized critical'' state for bacteria but instead find that the SOC sandpile critical behavior is very sensitive; critical behavior is destroyed with small perturbations effectively when the absorption of sand is introduced. It is only in the limit when the length of the genome of the bacteria tends to infinity that SOC properties are recovered for the asexual model.

  7. Driving risk assessment using near-crash database through data mining of tree-based model.

    PubMed

    Wang, Jianqiang; Zheng, Yang; Li, Xiaofei; Yu, Chenfei; Kodaka, Kenji; Li, Keqiang

    2015-11-01

    This paper considers a comprehensive naturalistic driving experiment to collect driving data under potential threats on actual Chinese roads. Using acquired real-world naturalistic driving data, a near-crash database is built, which contains vehicle status, potential crash objects, driving environment and road types, weather condition, and driver information and actions. The aims of this study are summarized into two aspects: (1) to cluster different driving-risk levels involved in near-crashes, and (2) to unveil the factors that greatly influence the driving-risk level. A novel method to quantify the driving-risk level of a near-crash scenario is proposed by clustering the braking process characteristics, namely maximum deceleration, average deceleration, and percentage reduction in vehicle kinetic energy. A classification and regression tree (CART) is employed to unveil the relationship among driving risk, driver/vehicle characteristics, and road environment. The results indicate that the velocity when braking, triggering factors, potential object type, and potential crash type exerted the greatest influence on the driving-risk levels in near-crashes.

  8. Research on an expert system for database operation of simulation-emulation math models. Volume 1, Phase 1: Results

    NASA Technical Reports Server (NTRS)

    Kawamura, K.; Beale, G. O.; Schaffer, J. D.; Hsieh, B. J.; Padalkar, S.; Rodriguez-Moscoso, J. J.

    1985-01-01

    The results of the first phase of Research on an Expert System for Database Operation of Simulation/Emulation Math Models, is described. Techniques from artificial intelligence (AI) were to bear on task domains of interest to NASA Marshall Space Flight Center. One such domain is simulation of spacecraft attitude control systems. Two related software systems were developed to and delivered to NASA. One was a generic simulation model for spacecraft attitude control, written in FORTRAN. The second was an expert system which understands the usage of a class of spacecraft attitude control simulation software and can assist the user in running the software. This NASA Expert Simulation System (NESS), written in LISP, contains general knowledge about digital simulation, specific knowledge about the simulation software, and self knowledge.

  9. ECMDB: the E. coli Metabolome Database.

    PubMed

    Guo, An Chi; Jewison, Timothy; Wilson, Michael; Liu, Yifeng; Knox, Craig; Djoumbou, Yannick; Lo, Patrick; Mandal, Rupasri; Krishnamurthy, Ram; Wishart, David S

    2013-01-01

    The Escherichia coli Metabolome Database (ECMDB, http://www.ecmdb.ca) is a comprehensively annotated metabolomic database containing detailed information about the metabolome of E. coli (K-12). Modelled closely on the Human and Yeast Metabolome Databases, the ECMDB contains >2600 metabolites with links to ∼1500 different genes and proteins, including enzymes and transporters. The information in the ECMDB has been collected from dozens of textbooks, journal articles and electronic databases. Each metabolite entry in the ECMDB contains an average of 75 separate data fields, including comprehensive compound descriptions, names and synonyms, chemical taxonomy, compound structural and physicochemical data, bacterial growth conditions and substrates, reactions, pathway information, enzyme data, gene/protein sequence data and numerous hyperlinks to images, references and other public databases. The ECMDB also includes an extensive collection of intracellular metabolite concentration data compiled from our own work as well as other published metabolomic studies. This information is further supplemented with thousands of fully assigned reference nuclear magnetic resonance and mass spectrometry spectra obtained from pure E. coli metabolites that we (and others) have collected. Extensive searching, relational querying and data browsing tools are also provided that support text, chemical structure, spectral, molecular weight and gene/protein sequence queries. Because of E. coli's importance as a model organism for biologists and as a biofactory for industry, we believe this kind of database could have considerable appeal not only to metabolomics researchers but also to molecular biologists, systems biologists and individuals in the biotechnology industry.

  10. Mechanism for production of secondary organic aerosols and their representation in atmospheric models. Final report

    SciTech Connect

    Seinfeld, J.H.; Flagan, R.C.

    1999-06-07

    This document contains the following: organic aerosol formation from the oxidation of biogenic hydrocarbons; gas/particle partitioning of semivolatile organic compounds to model inorganic, organic, and ambient smog aerosols; and representation of secondary organic aerosol formation in atmospheric models.

  11. A Comparative Data-Based Modeling Study on Respiratory CO2 Gas Exchange during Mechanical Ventilation

    PubMed Central

    Kim, Chang-Sei; Ansermino, J. Mark; Hahn, Jin-Oh

    2016-01-01

    The goal of this study is to derive a minimally complex but credible model of respiratory CO2 gas exchange that may be used in systematic design and pilot testing of closed-loop end-tidal CO2 controllers in mechanical ventilation. We first derived a candidate model that captures the essential mechanisms involved in the respiratory CO2 gas exchange process. Then, we simplified the candidate model to derive two lower-order candidate models. We compared these candidate models for predictive capability and reliability using experimental data collected from 25 pediatric subjects undergoing dynamically varying mechanical ventilation during surgical procedures. A two-compartment model equipped with transport delay to account for CO2 delivery between the lungs and the tissues showed modest but statistically significant improvement in predictive capability over the same model without transport delay. Aggregating the lungs and the tissues into a single compartment further degraded the predictive fidelity of the model. In addition, the model equipped with transport delay demonstrated superior reliability to the one without transport delay. Further, the respiratory parameters derived from the model equipped with transport delay, but not the one without transport delay, were physiologically plausible. The results suggest that gas transport between the lungs and the tissues must be taken into account to accurately reproduce the respiratory CO2 gas exchange process under conditions of wide-ranging and dynamically varying mechanical ventilation conditions. PMID:26870728

  12. Evaluation of a four-decade pan-European database of surface precipitation for river flow modeling

    NASA Astrophysics Data System (ADS)

    Gouweleeuw, Ben T.; Thielen, Jutta; de Roo, Ad; Cloke, Hannah; van der Knijff, Johan; Franchello, Gian

    2004-10-01

    The ECMWF Re Analysis (ERA-40) refers to the rerun of the European Centre of Medium Range Weather Forecast (ECMWF) Numerical Weather Prediction (NWP) model for the period September 1957- August 2002 employing all state-of-the-art information and satellite data input presently available. A selection of the reanalysis atmospheric output data can potentially be used to run a hydrological model to simulate historic river flows for the whole of Europe. Once evaluated against observed time series of rainfall and river flow, the output would constitute an extensive and coherent 40+ year database of pan-European calibrated river flow time series, providing a wealth of information and allowing a range of evaluation possibilities. Here, in order to separate the meteorological model performance from the hydrological model performance, the ERA-40 near-surface rainfall aggregates, which come as a by-product of the ECMWF NWP system, are evaluated against interpolated fields of observed surface precipitation. The ERA-40 rainfall fields consist of forecast data with a 36 hr lead time at midday and midnight and a 6 hr lead time at 6:00 and 18:00 UTC, allowing different combinations of lead and base times to compute daily rainfall aggregates. The evaluation of these aggregated precipitation fields against observed totals is relevant to the spin-up time of the ECMWF NWP system and the forecast reliability with increasing lead time. Interpolated fields of observed daily rainfall totals are provided by the Monitoring Agriculture with Remote Sensing (MARS) database (1990-2001) based at the European Commission Joint Research Centre (JRC).

  13. Model Organisms Fact Sheet: Using Model Organisms to Study Health and Disease

    MedlinePlus

    ... for the related ones, as well. What about computer models? Computers serve as virtual laboratories where scientists ... scientists have more confidence in the predictions. Can computer models replace animal models in research? Even though ...

  14. The expanding epigenetic landscape of non-model organisms.

    PubMed

    Bonasio, Roberto

    2015-01-01

    Epigenetics studies the emergence of different phenotypes from a single genotype. Although these processes are essential to cellular differentiation and transcriptional memory, they are also widely used in all branches of the tree of life by organisms that require plastic but stable adaptation to their physical and social environment. Because of the inherent flexibility of epigenetic regulation, a variety of biological phenomena can be traced back to evolutionary adaptations of few conserved molecular pathways that converge on chromatin. For these reasons chromatin biology and epigenetic research have a rich history of chasing discoveries in a variety of model organisms, including yeast, flies, plants and humans. Many more fascinating examples of epigenetic plasticity lie outside the realm of model organisms and have so far been only sporadically investigated at a molecular level; however, recent progress on sequencing technology and genome editing tools have begun to blur the lines between model and non-model organisms, opening numerous new avenues for investigation. Here, I review examples of epigenetic phenomena in non-model organisms that have emerged as potential experimental systems, including social insects, fish and flatworms, and are becoming accessible to molecular approaches.

  15. The expanding epigenetic landscape of non-model organisms

    PubMed Central

    Bonasio, Roberto

    2015-01-01

    Epigenetics studies the emergence of different phenotypes from a single genotype. Although these processes are essential to cellular differentiation and transcriptional memory, they are also widely used in all branches of the tree of life by organisms that require plastic but stable adaptation to their physical and social environment. Because of the inherent flexibility of epigenetic regulation, a variety of biological phenomena can be traced back to evolutionary adaptations of few conserved molecular pathways that converge on chromatin. For these reasons chromatin biology and epigenetic research have a rich history of chasing discoveries in a variety of model organisms, including yeast, flies, plants and humans. Many more fascinating examples of epigenetic plasticity lie outside the realm of model organisms and have so far been only sporadically investigated at a molecular level; however, recent progress on sequencing technology and genome editing tools have begun to blur the lines between model and non-model organisms, opening numerous new avenues for investigation. Here, I review examples of epigenetic phenomena in non-model organisms that have emerged as potential experimental systems, including social insects, fish and flatworms, and are becoming accessible to molecular approaches. PMID:25568458

  16. Making Organisms Model Human Behavior: Situated Models in North-American Alcohol Research, 1950-onwards

    PubMed Central

    Leonelli, Sabina; Ankeny, Rachel A.; Nelson, Nicole C.; Ramsden, Edmund

    2014-01-01

    Argument We examine the criteria used to validate the use of nonhuman organisms in North-American alcohol addiction research from the 1950s to the present day. We argue that this field, where the similarities between behaviors in humans and non-humans are particularly difficult to assess, has addressed questions of model validity by transforming the situatedness of non-human organisms into an experimental tool. We demonstrate that model validity does not hinge on the standardization of one type of organism in isolation, as often the case with genetic model organisms. Rather, organisms are viewed as necessarily situated: they cannot be understood as a model for human behavior in isolation from their environmental conditions. Hence the environment itself is standardized as part of the modeling process; and model validity is assessed with reference to the environmental conditions under which organisms are studied. PMID:25233743

  17. Lotka-Volterra competition models for sessile organisms.

    PubMed

    Spencer, Matthew; Tanner, Jason E

    2008-04-01

    Markov models are widely used to describe the dynamics of communities of sessile organisms, because they are easily fitted to field data and provide a rich set of analytical tools. In typical ecological applications, at any point in time, each point in space is in one of a finite set of states (e.g., species, empty space). The models aim to describe the probabilities of transitions between states. In most Markov models for communities, these transition probabilities are assumed to be independent of state abundances. This assumption is often suspected to be false and is rarely justified explicitly. Here, we start with simple assumptions about the interactions among sessile organisms and derive a model in which transition probabilities depend on the abundance of destination states. This model is formulated in continuous time and is equivalent to a Lotka-Volterra competition model. We fit this model and a variety of alternatives in which transition probabilities do not depend on state abundances to a long-term coral reef data set. The Lotka-Volterra model describes the data much better than all models we consider other than a saturated model (a model with a separate parameter for each transition at each time interval, which by definition fits the data perfectly). Our approach provides a basis for further development of stochastic models of sessile communities, and many of the methods we use are relevant to other types of community. We discuss possible extensions to spatially explicit models.

  18. A geographic information system on the potential distribution and abundance of Fasciola hepatica and F. gigantica in east Africa based on Food and Agriculture Organization databases.

    PubMed

    Malone, J B; Gommes, R; Hansen, J; Yilma, J M; Slingenberg, J; Snijders, F; Nachtergaele, F; Ataman, E

    1998-07-31

    An adaptation of a previously developed climate forecast computer model and digital agroecologic database resources available from FAO for developing countries were used to develop a geographic information system risk assessment model for fasciolosis in East Africa, a region where both F. hepatica and F. gigantica occur as a cause of major economic losses in livestock. Regional F. hepatica and F. gigantica forecast index maps were created. Results were compared to environmental data parameters, known life cycle micro-environment requirements and to available Fasciola prevalence survey data and distribution patterns reported in the literature for each species (F. hepatica above 1200 m elevation, F. gigantica below 1800 m, both at 1200-1800 m). The greatest risk, for both species, occurred in areas of extended high annual rainfall associated with high soil moisture and surplus water, with risk diminishing in areas of shorter wet season and/or lower temperatures. Arid areas were generally unsuitable (except where irrigation, water bodies or floods occur) due to soil moisture deficit and/or, in the case of F. hepatica, high average annual mean temperature >23 degrees C. Regions in the highlands of Ethiopia and Kenya were identified as unsuitable for F. gigantica due to inadequate thermal regime, below the 600 growing degree days required for completion of the life cycle in a single year. The combined forecast index (F. hepatica+F. gigantica) was significantly correlated to prevalence data available for 260 of the 1220 agroecologic crop production system zones (CPSZ) and to average monthly normalized difference vegetation index (NDVI) values derived from the advanced very high resolution radiometer (AVHRR) sensor on board the NOAA polar-orbiting satellites. For use in Fasciola control programs, results indicate that monthly forecast parameters, developed in a GIS with digital agroecologic zone databases and monthly climate databases, can be used to define the

  19. Flow, mixing, and displacement in using a data-based hydrochemical model to predict conservative tracer data

    NASA Astrophysics Data System (ADS)

    Iorgulescu, I.; Beven, K. J.; Musy, A.

    2007-03-01

    We extend the data-based hydrochemical model of Iorgulescu et al. (2005), able to simulate discharge and reactive chemical tracer concentrations (silica and calcium) in streamflow for subcatchments of the Haute-Mentue research basin (Switzerland), to the prediction of additional δ18O values treated as a conservative tracer. The hydrochemical model is based on a parameterization of three runoff components (direct precipitation (DP), acid soil water (AS), and deep groundwater (GW)) in a chemical mixing model. Each component is modeled through an identical structure consisting of a nonlinear gain and a linear transfer function with two reservoirs (fast/slow) in parallel having a constant partition between them. We formulate a set of hypotheses concerning the isotope characterization of each component to provide additional information about how new rainfall inputs are processed in the hydrological response of the catchment. In particular, the AS component is modeled through a nested structure of hypotheses (models) of increasing complexity. It will be shown that hydrological processes in the hillslope associated with the DP, AS, and GW components are especially effective in filtering of higher-frequency fluctuations in precipitation isotopic ratios at the intraevent, interevent/seasonal, and annual/multiannual timescales. The highly nonlinear and nonstationary AS component represents predominantly "recent" water stored in the upper decimeters of the soil profile. Results also suggest that subsurface pathways are significant for the DP component. A local flow path mechanism is proposed for explaining the large fluxes of subsurface flows.

  20. European Community Databases: Online to Europe.

    ERIC Educational Resources Information Center

    Hensley, Colin

    1989-01-01

    Describes three groups of databases sponsored by the European Communities Commission: Eurobases, a textual database of the contents of the "Official Journal" of the European Community; the European Community Host Organization (ECHO) databases, which offer multilingual information about Europe; and statistical databases. Information on access and…

  1. MOAtox: A Comprehensive Mode of Action and Acute Aquatic Toxicity Database for Predictive Model Development

    EPA Science Inventory

    tThe mode of toxic action (MOA) has been recognized as a key determinant of chemical toxicity andas an alternative to chemical class-based predictive toxicity modeling. However, the development ofquantitative structure activity relationship (QSAR) and other models has been limite...

  2. Millennial Students' Mental Models of Search: Implications for Academic Librarians and Database Developers

    ERIC Educational Resources Information Center

    Holman, Lucy

    2011-01-01

    Today's students exhibit generational differences in the way they search for information. Observations of first-year students revealed a proclivity for simple keyword or phrases searches with frequent misspellings and incorrect logic. Although no students had strong mental models of search mechanisms, those with stronger models did construct more…

  3. Modeling aspects of estuarine eutrophication. (Latest citations from the Selected Water Resources Abstracts database). Published Search

    SciTech Connect

    Not Available

    1993-05-01

    The bibliography contains citations concerning mathematical modeling of existing water quality stresses in estuaries, harbors, bays, and coves. Both physical hydraulic and numerical models for estuarine circulation are discussed. (Contains a minimum of 96 citations and includes a subject term index and title list.)

  4. Modelling of organic matter dynamics during the composting process.

    PubMed

    Zhang, Y; Lashermes, G; Houot, S; Doublet, J; Steyer, J P; Zhu, Y G; Barriuso, E; Garnier, P

    2012-01-01

    Composting urban organic wastes enables the recycling of their organic fraction in agriculture. The objective of this new composting model was to gain a clearer understanding of the dynamics of organic fractions during composting and to predict the final quality of composts. Organic matter was split into different compartments according to its degradability. The nature and size of these compartments were studied using a biochemical fractionation method. The evolution of each compartment and the microbial biomass were simulated, as was the total organic carbon loss corresponding to organic carbon mineralisation into CO(2). Twelve composting experiments from different feedstocks were used to calibrate and validate our model. We obtained a unique set of estimated parameters. Good agreement was achieved between the simulated and experimental results that described the evolution of different organic fractions, with the exception of some compost because of a poor simulation of the cellulosic and soluble pools. The degradation rate of the cellulosic fraction appeared to be highly variable and dependent on the origin of the feedstocks. The initial soluble fraction could contain some degradable and recalcitrant elements that are not easily accessible experimentally. PMID:21978424

  5. Experimental study on a general data-based nonlinear identification approach for a MDOF frame model structure

    NASA Astrophysics Data System (ADS)

    Xu, Bin; He, Jia; Zhou, Ren; Masri, Sami F.

    2010-05-01

    Though many system identification methods are currently available for structural physical parameter estimation, a major challenge lies in obtaining reasonably accurate identification results for nonlinear structural systems because the initiation and development of structural damage are typical nonlinear process. Moreover, in most current model updating and identification approaches, structural mass distribution is usually treated as known. But the mass of an engineering structure can not be estimated correctly according to the design drawing directly. In this study, a general data-based approach using structural dynamic response and excitation measurement time series was proposed and validated for the nonlinear behavior identification of a 4-story frame structure with smart devices of magneto-rheological (MR) damper to simulate nonlinear performance. The whole approach includes a two-stage procedure. In the first stage, least-squares technique was used to identify the system matrices of an equivalent linear system basing on the impact force applied on each floor of the structure and the corresponding vibration measurements; and in the second stage, the nonlinear restoring force of the MR damper was identified and compared with the measurement with a force gauge. The performance of the proposed approach was studied when current level of 0.15A for the MR damper was employed. Results show that the proposed data-based approach is capable of identifying nonlinear behavior of engineering structures.

  6. An Extensible, Interchangeable and Sharable Database Model for Improving Multidisciplinary Aircraft Design

    NASA Technical Reports Server (NTRS)

    Lin, Risheng; Afjeh, Abdollah A.

    2003-01-01

    Crucial to an efficient aircraft simulation-based design is a robust data modeling methodology for both recording the information and providing data transfer readily and reliably. To meet this goal, data modeling issues involved in the aircraft multidisciplinary design are first analyzed in this study. Next, an XML-based. extensible data object model for multidisciplinary aircraft design is constructed and implemented. The implementation of the model through aircraft databinding allows the design applications to access and manipulate any disciplinary data with a lightweight and easy-to-use API. In addition, language independent representation of aircraft disciplinary data in the model fosters interoperability amongst heterogeneous systems thereby facilitating data sharing and exchange between various design tools and systems.

  7. Precisely parameterized experimental and computational models of tissue organization.

    PubMed

    Molitoris, Jared M; Paliwal, Saurabh; Sekar, Rajesh B; Blake, Robert; Park, JinSeok; Trayanova, Natalia A; Tung, Leslie; Levchenko, Andre

    2016-02-01

    Patterns of cellular organization in diverse tissues frequently display a complex geometry and topology tightly related to the tissue function. Progressive disorganization of tissue morphology can lead to pathologic remodeling, necessitating the development of experimental and theoretical methods of analysis of the tolerance of normal tissue function to structural alterations. A systematic way to investigate the relationship of diverse cell organization to tissue function is to engineer two-dimensional cell monolayers replicating key aspects of the in vivo tissue architecture. However, it is still not clear how this can be accomplished on a tissue level scale in a parameterized fashion, allowing for a mathematically precise definition of the model tissue organization and properties down to a cellular scale with a parameter dependent gradual change in model tissue organization. Here, we describe and use a method of designing precisely parameterized, geometrically complex patterns that are then used to control cell alignment and communication of model tissues. We demonstrate direct application of this method to guiding the growth of cardiac cell cultures and developing mathematical models of cell function that correspond to the underlying experimental patterns. Several anisotropic patterned cultures spanning a broad range of multicellular organization, mimicking the cardiac tissue organization of different regions of the heart, were found to be similar to each other and to isotropic cell monolayers in terms of local cell-cell interactions, reflected in similar confluency, morphology and connexin-43 expression. However, in agreement with the model predictions, different anisotropic patterns of cell organization, paralleling in vivo alterations of cardiac tissue morphology, resulted in variable and novel functional responses with important implications for the initiation and maintenance of cardiac arrhythmias. We conclude that variations of tissue geometry and topology

  8. Modelling the fate of oxidisable organic contaminants in groundwater

    NASA Astrophysics Data System (ADS)

    Barry, D. A.; Prommer, H.; Miller, C. T.; Engesgaard, P.; Brun, A.; Zheng, C.

    Subsurface contamination by organic chemicals is a pervasive environmental problem, susceptible to remediation by natural or enhanced attenuation approaches or more highly engineered methods such as pump-and-treat, amongst others. Such remediation approaches, along with risk assessment or the pressing need to address complex scientific questions, have driven the development of integrated modelling tools that incorporate physical, biological and geochemical processes. We provide a comprehensive modelling framework, including geochemical reactions and interphase mass transfer processes such as sorption/desorption, non-aqueous phase liquid dissolution and mineral precipitatation/dissolution, all of which can be in equilibrium or kinetically controlled. This framework is used to simulate microbially mediated transformation/degradation processes and the attendant microbial population growth and decay. Solution algorithms, particularly the split-operator (SO) approach, are described, along with a brief résumé of numerical solution methods. Some of the available numerical models are described, mainly those constructed using available flow, transport and geochemical reaction packages. The general modelling framework is illustrated by pertinent examples, showing the degradation of dissolved organics by microbial activity limited by the availability of nutrients or electron acceptors (i.e., changing redox states), as well as concomitant secondary reactions. Two field-scale modelling examples are discussed, the Vejen landfill (Denmark) and an example where metal contamination is remediated by redox changes wrought by injection of a dissolved organic compound. A summary is provided of current and likely future challenges to modelling of oxidisable organics in the subsurface.

  9. Electrochemical model of the polyaniline based organic memristive device

    SciTech Connect

    Demin, V. A. E-mail: victor.erokhin@fis.unipr.it; Erokhin, V. V. E-mail: victor.erokhin@fis.unipr.it; Kashkarov, P. K.; Kovalchuk, M. V.

    2014-08-14

    The electrochemical organic memristive device with polyaniline active layer is a stand-alone device designed and realized for reproduction of some synapse properties in the innovative electronic circuits, including the neuromorphic networks capable for learning. In this work, a new theoretical model of the polyaniline memristive is presented. The developed model of organic memristive functioning was based on the detailed consideration of possible electrochemical processes occuring in the active zone of this device. Results of the calculation have demonstrated not only the qualitative explanation of the characteristics observed in the experiment but also the quantitative similarities of the resultant current values. It is shown how the memristive could behave at zero potential difference relative to the reference electrode. This improved model can establish a basis for the design and prediction of properties of more complicated circuits and systems (including stochastic ones) based on the organic memristive devices.

  10. Enhancing medical database semantics.

    PubMed Central

    Leão, B. de F.; Pavan, A.

    1995-01-01

    Medical Databases deal with dynamic, heterogeneous and fuzzy data. The modeling of such complex domain demands powerful semantic data modeling methodologies. This paper describes GSM-Explorer a Case Tool that allows for the creation of relational databases using semantic data modeling techniques. GSM Explorer fully incorporates the Generic Semantic Data Model-GSM enabling knowledge engineers to model the application domain with the abstraction mechanisms of generalization/specialization, association and aggregation. The tool generates a structure that implements persistent database-objects through the automatic generation of customized SQL ANSI scripts that sustain the semantics defined in the higher lever. This paper emphasizes the system architecture and the mapping of the semantic model into relational tables. The present status of the project and its further developments are discussed in the Conclusions. PMID:8563288

  11. Construction and analysis of a human hepatotoxicity database suitable for QSAR modeling using post-market safety data.

    PubMed

    Zhu, Xiao; Kruhlak, Naomi L

    2014-07-01

    Drug-induced liver injury (DILI) is one of the most common drug-induced adverse events (AEs) leading to life-threatening conditions such as acute liver failure. It has also been recognized as the single most common cause of safety-related post-market withdrawals or warnings. Efforts to develop new predictive methods to assess the likelihood of a drug being a hepatotoxicant have been challenging due to the complexity and idiosyncrasy of clinical manifestations of DILI. The FDA adverse event reporting system (AERS) contains post-market data that depict the morbidity of AEs. Here, we developed a scalable approach to construct a hepatotoxicity database using post-market data for the purpose of quantitative structure-activity relationship (QSAR) modeling. A set of 2029 unique and modelable drug entities with 13,555 drug-AE combinations was extracted from the AERS database using 37 hepatotoxicity-related query preferred terms (PTs). In order to determine the optimal classification scheme to partition positive from negative drugs, a manually-curated DILI calibration set composed of 105 negatives and 177 positives was developed based on the published literature. The final classification scheme combines hepatotoxicity-related PT data with supporting information that optimize the predictive performance across the calibration set. Data for other toxicological endpoints related to liver injury such as liver enzyme abnormalities, cholestasis, and bile duct disorders, were also extracted and classified. Collectively, these datasets can be used to generate a battery of QSAR models that assess a drug's potential to cause DILI.

  12. Construction and analysis of a human hepatotoxicity database suitable for QSAR modeling using post-market safety data.

    PubMed

    Zhu, Xiao; Kruhlak, Naomi L

    2014-07-01

    Drug-induced liver injury (DILI) is one of the most common drug-induced adverse events (AEs) leading to life-threatening conditions such as acute liver failure. It has also been recognized as the single most common cause of safety-related post-market withdrawals or warnings. Efforts to develop new predictive methods to assess the likelihood of a drug being a hepatotoxicant have been challenging due to the complexity and idiosyncrasy of clinical manifestations of DILI. The FDA adverse event reporting system (AERS) contains post-market data that depict the morbidity of AEs. Here, we developed a scalable approach to construct a hepatotoxicity database using post-market data for the purpose of quantitative structure-activity relationship (QSAR) modeling. A set of 2029 unique and modelable drug entities with 13,555 drug-AE combinations was extracted from the AERS database using 37 hepatotoxicity-related query preferred terms (PTs). In order to determine the optimal classification scheme to partition positive from negative drugs, a manually-curated DILI calibration set composed of 105 negatives and 177 positives was developed based on the published literature. The final classification scheme combines hepatotoxicity-related PT data with supporting information that optimize the predictive performance across the calibration set. Data for other toxicological endpoints related to liver injury such as liver enzyme abnormalities, cholestasis, and bile duct disorders, were also extracted and classified. Collectively, these datasets can be used to generate a battery of QSAR models that assess a drug's potential to cause DILI. PMID:24721472

  13. Molecular analysis of the replication program in unicellular model organisms

    PubMed Central

    Brewer, Bonita J.

    2014-01-01

    Eukaryotes have long been reported to show temporal programs of replication, different portions of the genome being replicated at different times in S phase, with the added possibility of developmentally regulated changes in this pattern depending on species and cell type. Unicellular model organisms, primarily the budding yeast Saccharomyces cerevisiae, have been central to our current understanding of the mechanisms underlying the regulation of replication origins and the temporal program of replication in particular. But what exactly is a temporal program of replication, and how might it arise? In this article, we explore this question, drawing again on the wealth of experimental information in unicellular model organisms. PMID:20012185

  14. Modeling of Spatially Correlated Energetic Disorder in Organic Semiconductors.

    PubMed

    Kordt, Pascal; Andrienko, Denis

    2016-01-12

    Mesoscale modeling of organic semiconductors relies on solving an appropriately parametrized master equation. Essential ingredients of the parametrization are site energies (driving forces), which enter the charge transfer rate between pairs of neighboring molecules. Site energies are often Gaussian-distributed and are spatially correlated. Here, we propose an algorithm that generates these energies with a given Gaussian distribution and spatial correlation function. The method is tested on an amorphous organic semiconductor, DPBIC, illustrating that the accurate description of correlations is essential for the quantitative modeling of charge transport in amorphous mesophases. PMID:26588022

  15. Self-Organizing Neural Network Models for State Anticipatory Systems

    NASA Astrophysics Data System (ADS)

    Pöllä, Matti; Honkela, Timo

    2006-06-01

    A vital mechanism of high-level natural cognitive systems is the anticipatory capability of making decisions based on predicted events in the future. While in some cases the performance of computational cognitive systems can be improved by modeling anticipatory behavior, it has been shown that for many cognitive tasks anticipation is mandatory. In this paper, we review the use of self-organizing artificial neural networks in constructing the state-space model of an anticipatory system. The biologically inspired self-organizing map (SOM) and its topologically dynamic variants such as the growing neural gas (GNG) are discussed using illustrative examples of their performance.

  16. The National Institute on Disability and Rehabilitation Research burn model system database: a tool for the multicenter study of the outcome of burn injury.

    PubMed

    Klein, Matthew B; Lezotte, Dennis L; Fauerbach, James A; Herndon, David N; Kowalske, Karen J; Carrougher, Gretchen J; deLateur, Barbara J; Holavanahalli, Radha; Esselman, Peter C; San Agustin, Theresa B; Engrav, Loren H

    2007-01-01

    Advances in critical care and surgical management have significantly improved survival after burn injury over the past several decades. However, today, survival alone is an insufficient outcome. In 1994, the National Institute on Disability and Rehabilitation Research (NIDRR) created a burn model system program to evaluate the long-term sequelae of burn injuries. As part of this multicenter program, a comprehensive demographic and outcome database was developed to facilitate the study of a number of functional and psychosocial outcomes after burns. The purpose of this study is to review the database design and structure as well as the data obtained during the last 10 years. This is a descriptive study of the NIDRR database structure as well as the patient data obtained from the four participating burn centers from 1994 to 2004. Data obtained during hospitalization and at 6, 12, and 24 months after discharge were reviewed and descriptive statistics were calculated for select database fields. The database is divided into several subsections, including demographics, injury complications, patient disposition, and functional and psychological surveys. A total of 4600 patients have been entered into the NIDRR database. To date, 3449 (75%) patients were alive at discharged and consented to follow-up data collection. The NIDRR database provides an expansive repository of patient, injury, and outcome data that can be used to analyze the impact of burn injury on physical and psychosocial function and for the design of interventions to enhance the quality of life of burn survivors. PMID:17211206

  17. Porphyry Copper Deposits of the World: Database and Grade and Tonnage Models, 2008

    USGS Publications Warehouse

    Singer, Donald A.; Berger, Vladimir I.; Moring, Barry C.

    2008-01-01

    This report is an update of earlier publications about porphyry copper deposits (Singer, Berger, and Moring, 2002; Singer, D.A., Berger, V.I., and Moring, B.C., 2005). The update was necessary because of new information about substantial increases in resources in some deposits and because we revised locations of some deposits so that they are consistent with images in GoogleEarth. In this report we have added new porphyry copper deposits and removed a few incorrectly classed deposits. In addition, some errors have been corrected and a number of deposits have had some information, such as grades, tonnages, locations, or ages revised. Colleagues have helped identify places where improvements were needed. Mineral deposit models are important in exploration planning and quantitative resource assessments for a number of reasons including: (1) grades and tonnages among deposit types are significantly different, and (2) many types occur in different geologic settings that can be identified from geologic maps. Mineral deposit models are the keystone in combining the diverse geoscience information on geology, mineral occurrences, geophysics, and geochemistry used in resource assessments and mineral exploration. Too few thoroughly explored mineral deposits are available in most local areas for reliable identification of the important geoscience variables or for robust estimation of undiscovered deposits?thus we need mineral-deposit models. Globally based deposit models allow recognition of important features because the global models demonstrate how common different features are. Well-designed and -constructed deposit models allow geologists to know from observed geologic environments the possible mineral deposit types that might exist, and allow economists to determine the possible economic viability of these resources in the region. Thus, mineral deposit models play the central role in transforming geoscience information to a form useful to policy makers. The foundation of

  18. Sediment-Hosted Zinc-Lead Deposits of the World - Database and Grade and Tonnage Models

    USGS Publications Warehouse

    Singer, Donald A.; Berger, Vladimir I.; Moring, Barry C.

    2009-01-01

    This report provides information on sediment-hosted zinc-lead mineral deposits based on the geologic settings that are observed on regional geologic maps. The foundation of mineral-deposit models is information about known deposits. The purpose of this publication is to make this kind of information available in digital form for sediment-hosted zinc-lead deposits. Mineral-deposit models are important in exploration planning and quantitative resource assessments: Grades and tonnages among deposit types are significantly different, and many types occur in different geologic settings that can be identified from geologic maps. Mineral-deposit models are the keystone in combining the diverse geoscience information on geology, mineral occurrences, geophysics, and geochemistry used in resource assessments and mineral exploration. Too few thoroughly explored mineral deposits are available in most local areas for reliable identification of the important geoscience variables, or for robust estimation of undiscovered deposits - thus, we need mineral-deposit models. Globally based deposit models allow recognition of important features because the global models demonstrate how common different features are. Well-designed and -constructed deposit models allow geologists to know from observed geologic environments the possible mineral-deposit types that might exist, and allow economists to determine the possible economic viability of these resources in the region. Thus, mineral-deposit models play the central role in transforming geoscience information to a form useful to policy makers. This publication contains a computer file of information on sediment-hosted zinc-lead deposits from around the world. It also presents new grade and tonnage models for nine types of these deposits and a file allowing locations of all deposits to be plotted in Google Earth. The data are presented in FileMaker Pro, Excel and text files to make the information available to as many as possible. The

  19. Porphyry copper deposits of the world: database, map, and grade and tonnage models

    USGS Publications Warehouse

    Singer, Donald A.; Berger, Vladimir Iosifovich; Moring, Barry C.

    2005-01-01

    Mineral deposit models are important in exploration planning and quantitative resource assessments for two reasons: (1) grades and tonnages among deposit types are significantly different, and (2) many types occur in different geologic settings that can be identified from geologic maps. Mineral deposit models are the keystone in combining the diverse geoscience information on geology, mineral occurrences, geophysics, and geochemistry used in resource assessments and mineral exploration. Too few thoroughly explored mineral deposits are available in most local areas for reliable identification of the important geoscience variables or for robust estimation of undiscovered deposits-thus we need mineral-deposit models. Globally based deposit models allow recognition of important features because the global models demonstrate how common different features are. Well-designed and -constructed deposit models allow geologists to know from observed geologic environments the possible mineral deposit types that might exist, and allow economists to determine the possible economic viability of these resources in the region. Thus, mineral deposit models play the central role in transforming geoscience information to a form useful to policy makers. The foundation of mineral deposit models is information about known deposits-the purpose of this publication is to make this kind of information available in digital form for porphyry copper deposits. This report is an update of an earlier publication about porphyry copper deposits. In this report we have added 84 new porphyry copper deposits and removed 12 deposits. In addition, some errors have been corrected and a number of deposits have had some information, such as grades, tonnages, locations, or ages revised. This publication contains a computer file of information on porphyry copper deposits from around the world. It also presents new grade and tonnage models for porphyry copper deposits and for three subtypes of porphyry copper

  20. The MIntAct Project and Molecular Interaction Databases.

    PubMed

    Licata, Luana; Orchard, Sandra

    2016-01-01

    Molecular interaction databases collect, organize, and enable the analysis of the increasing amounts of molecular interaction data being produced and published as we move towards a more complete understanding of the interactomes of key model organisms. The organization of these data in a structured format supports analyses such as the modeling of pairwise relationships between interactors into interaction networks and is a powerful tool for understanding the complex molecular machinery of the cell. This chapter gives an overview of the principal molecular interaction databases, in particular the IMEx databases, and their curation policies, use of standardized data formats and quality control rules. Special attention is given to the MIntAct project, in which IntAct and MINT joined forces to create a single resource to improve curation and software development efforts. This is exemplified as a model for the future of molecular interaction data collation and dissemination. PMID:27115627

  1. Prediction of organ toxicity endpoints by QSAR modeling based on precise chemical-histopathology annotations.

    PubMed

    Myshkin, Eugene; Brennan, Richard; Khasanova, Tatiana; Sitnik, Tatiana; Serebriyskaya, Tatiana; Litvinova, Elena; Guryanov, Alexey; Nikolsky, Yuri; Nikolskaya, Tatiana; Bureeva, Svetlana

    2012-09-01

    The ability to accurately predict the toxicity of drug candidates from their chemical structure is critical for guiding experimental drug discovery toward safer medicines. Under the guidance of the MetaTox consortium (Thomson Reuters, CA, USA), which comprised toxicologists from the pharmaceutical industry and government agencies, we created a comprehensive ontology of toxic pathologies for 19 organs, classifying pathology terms by pathology type and functional organ substructure. By manual annotation of full-text research articles, the ontology was populated with chemical compounds causing specific histopathologies. Annotated compound-toxicity associations defined histologically from rat and mouse experiments were used to build quantitative structure-activity relationship models predicting subcategories of liver and kidney toxicity: liver necrosis, liver relative weight gain, liver lipid accumulation, nephron injury, kidney relative weight gain, and kidney necrosis. All models were validated using two independent test sets and demonstrated overall good performance: initial validation showed 0.80-0.96 sensitivity (correctly predicted toxic compounds) and 0.85-1.00 specificity (correctly predicted non-toxic compounds). Later validation against a test set of compounds newly added to the database in the 2 years following initial model generation showed 75-87% sensitivity and 60-78% specificity. General hepatotoxicity and nephrotoxicity models were less accurate, as expected for more complex endpoints.

  2. EPIC modeling of soil organic carbon sequestration in croplands of Iowa.

    PubMed

    Causarano, Hector J; Doraiswamy, Paul C; McCarty, Gregory W; Hatfield, Jerry L; Milak, Sushil; Stern, Alan J

    2008-01-01

    Depending on management, soil organic carbon (SOC) is a potential source or sink for atmospheric CO(2). We used the EPIC model to study impacts of soil and crop management on SOC in corn (Zea mays L.) and soybean (Glycine max L. Merr.) croplands of Iowa. The National Agricultural Statistics Service crops classification maps were used to identify corn-soybean areas. Soil properties were obtained from a combination of SSURGO and STATSGO databases. Daily weather variables were obtained from first order meteorological stations in Iowa and neighboring states. Data on crop management, fertilizer application and tillage were obtained from publicly available databases maintained by the NRCS, USDA-Economic Research Service (ERS), and Conservation Technology Information Center. The EPIC model accurately simulated state averages of crop yields during 1970-2005 (R(2) = 0.87). Simulated SOC explained 75% of the variation in measured SOC. With current trends in conservation tillage adoption, total stock of SOC (0-20 cm) is predicted to reach 506 Tg by 2019, representing an increase of 28 Tg with respect to 1980. In contrast, when the whole soil profile was considered, EPIC estimated a decrease of SOC stocks with time, from 1835 Tg in 1980 to 1771 Tg in 2019. Hence, soil depth considered for calculations is an important factor that needs further investigation. Soil organic C sequestration rates (0-20 cm) were estimated at 0.50 to 0.63 Mg ha(-1) yr(-1) depending on climate and soil conditions. Overall, combining land use maps with EPIC proved valid for predicting impacts of management practices on SOC. However, more data on spatial and temporal variation in SOC are needed to improve model calibration and validation.

  3. The NorWeST Stream Temperature Database, Model, and Climate Scenarios for the Northwest U.S. (Invited)

    NASA Astrophysics Data System (ADS)

    Isaak, D.; Wenger, S.; Peterson, E.; Ver Hoef, J.; Luce, C.; Hostetler, S. W.; Kershner, J.; Dunham, J.; Nagel, D.; Roper, B.

    2013-12-01

    Anthropogenic climate change is warming the Earth's rivers and streams and threatens significant changes to aquatic biodiversity. Effective threat response will require prioritization of limited conservation resources and coordinated interagency efforts guided by accurate information about climate, and climate change, at scales relevant to the distributions of species across landscapes. Here, we describe the NorWeST (i.e., NorthWest Stream Temperature) project to develop a comprehensive interagency stream temperature database and high-resolution climate scenarios across Washington, Oregon, Idaho, Montana, and Wyoming (~400,000 stream kilometers). The NorWeST database consists of stream temperature data contributed by >60 state, federal, tribal, and private resource agencies and may be the largest of its kind in the world (>45,000,000 hourly temperature recordings at >15,000 unique monitoring sites). These data are being used with spatial statistical network models to accurately downscale (R2 = 90%; RMSE < 1 C) global climate patterns to all perennially flowing reaches within river networks at 1-kilometer resolution. Historic stream temperature scenarios are developed using air temperature data from RegCM3 runs for the NCEP historical reanalysis and future scenarios (2040s and 2080s) are developed by applying bias corrected air temperature and discharge anomalies from ensemble climate and hydrology model runs for A1B and A2 warming trajectories. At present, stream temperature climate scenarios have been developed for 230,000 stream kilometers across Idaho and western Montana using data from more than 7,000 monitoring sites. The raw temperature data and stream climate scenarios are made available as ArcGIS geospatial products for download through the NorWeST website as individual river basins are completed (http://www.fs.fed.us/rm/boise/AWAE/projects/NorWeST.shtml). By providing open access to temperature data and scenarios, the project is fostering new research on

  4. Workshop meeting report Organs-on-Chips: human disease models.

    PubMed

    van de Stolpe, Anja; den Toonder, Jaap

    2013-09-21

    The concept of "Organs-on-Chips" has recently evolved and has been described as 3D (mini-) organs or tissues consisting of multiple and different cell types interacting with each other under closely controlled conditions, grown in a microfluidic chip, and mimicking the complex structures and cellular interactions in and between different cell types and organs in vivo, enabling the real time monitoring of cellular processes. In combination with the emerging iPSC (induced pluripotent stem cell) field this development offers unprecedented opportunities to develop human in vitro models for healthy and diseased organ tissues, enabling the investigation of fundamental mechanisms in disease development, drug toxicity screening, drug target discovery and drug development, and the replacement of animal testing. Capturing the genetic background of the iPSC donor in the organ or disease model carries the promise to move towards "in vitro clinical trials", reducing costs for drug development and furthering the concept of personalized medicine and companion diagnostics. During the Lorentz workshop (Leiden, September 2012) an international multidisciplinary group of experts discussed the current state of the art, available and emerging technologies, applications and how to proceed in the field. Organ-on-a-chip platform technologies are expected to revolutionize cell biology in general and drug development in particular. PMID:23645172

  5. Transferable Atomic Multipole Machine Learning Models for Small Organic Molecules.

    PubMed

    Bereau, Tristan; Andrienko, Denis; von Lilienfeld, O Anatole

    2015-07-14

    Accurate representation of the molecular electrostatic potential, which is often expanded in distributed multipole moments, is crucial for an efficient evaluation of intermolecular interactions. Here we introduce a machine learning model for multipole coefficients of atom types H, C, O, N, S, F, and Cl in any molecular conformation. The model is trained on quantum-chemical results for atoms in varying chemical environments drawn from thousands of organic molecules. Multipoles in systems with neutral, cationic, and anionic molecular charge states are treated with individual models. The models' predictive accuracy and applicability are illustrated by evaluating intermolecular interaction energies of nearly 1,000 dimers and the cohesive energy of the benzene crystal.

  6. Review of existing terrestrial bioaccumulation models and terrestrial bioaccumulation modeling needs for organic chemicals

    EPA Science Inventory

    Protocols for terrestrial bioaccumulation assessments are far less-developed than for aquatic systems. This manuscript reviews modeling approaches that can be used to assess the terrestrial bioaccumulation potential of commercial organic chemicals. Models exist for plant, inver...

  7. Implementing marine organic aerosols into the GEOS-Chem model

    DOE PAGESBeta

    Gantt, B.; Johnson, M. S.; Crippa, M.; Prévôt, A. S. H.; Meskhidze, N.

    2014-09-09

    Marine organic aerosols (MOA) have been shown to play an important role in tropospheric chemistry by impacting surface mass, cloud condensation nuclei, and ice nuclei concentrations over remote marine and coastal regions. In this work, an online marine primary organic aerosol emission parameterization, designed to be used for both global and regional models, was implemented into the GEOS-Chem model. The implemented emission scheme improved the large underprediction of organic aerosol concentrations in clean marine regions (normalized mean bias decreases from -79% when using the default settings to -12% when marine organic aerosols are added). Model predictions were also in goodmore » agreement (correlation coefficient of 0.62 and normalized mean bias of -36%) with hourly surface concentrations of MOA observed during the summertime at an inland site near Paris, France. Our study shows that MOA have weaker coastal-to-inland concentration gradients than sea-salt aerosols, leading to several inland European cities having > 10% of their surface submicron organic aerosol mass concentration with a marine source. The addition of MOA tracers to GEOS-Chem enabled us to identify the regions with large contributions of freshly-emitted or aged aerosol having distinct physicochemical properties, potentially indicating optimal locations for future field studies.« less

  8. Implementing marine organic aerosols into the GEOS-Chem model

    NASA Astrophysics Data System (ADS)

    Gantt, B.; Johnson, M. S.; Crippa, M.; Prévôt, A. S. H.; Meskhidze, N.

    2015-03-01

    Marine-sourced organic aerosols (MOAs) have been shown to play an important role in tropospheric chemistry by impacting surface mass, cloud condensation nuclei, and ice nuclei concentrations over remote marine and coastal regions. In this work, an online marine primary organic aerosol emission parameterization, designed to be used for both global and regional models, was implemented into the GEOS-Chem (Global Earth Observing System Chemistry) model. The implemented emission scheme improved the large underprediction of organic aerosol concentrations in clean marine regions (normalized mean bias decreases from -79% when using the default settings to -12% when marine organic aerosols are added). Model predictions were also in good agreement (correlation coefficient of 0.62 and normalized mean bias of -36%) with hourly surface concentrations of MOAs observed during the summertime at an inland site near Paris, France. Our study shows that MOAs have weaker coastal-to-inland concentration gradients than sea-salt aerosols, leading to several inland European cities having >10% of their surface submicron organic aerosol mass concentration with a marine source. The addition of MOA tracers to GEOS-Chem enabled us to identify the regions with large contributions of freshly emitted or aged aerosol having distinct physicochemical properties, potentially indicating optimal locations for future field studies.

  9. Implementing marine organic aerosols into the GEOS-Chem model

    NASA Astrophysics Data System (ADS)

    Gantt, B.; Johnson, M. S.; Crippa, M.; Prévôt, A. S. H.; Meskhidze, N.

    2014-09-01

    Marine organic aerosols (MOA) have been shown to play an important role in tropospheric chemistry by impacting surface mass, cloud condensation nuclei, and ice nuclei concentrations over remote marine and coastal regions. In this work, an online marine primary organic aerosol emission parameterization, designed to be used for both global and regional models, was implemented into the GEOS-Chem model. The implemented emission scheme improved the large underprediction of organic aerosol concentrations in clean marine regions (normalized mean bias decreases from -79% when using the default settings to -12% when marine organic aerosols are added). Model predictions were also in good agreement (correlation coefficient of 0.62 and normalized mean bias of -36%) with hourly surface concentrations of MOA observed during the summertime at an inland site near Paris, France. Our study shows that MOA have weaker coastal-to-inland concentration gradients than sea-salt aerosols, leading to several inland European cities having > 10% of their surface submicron organic aerosol mass concentration with a marine source. The addition of MOA tracers to GEOS-Chem enabled us to identify the regions with large contributions of freshly-emitted or aged aerosol having distinct physicochemical properties, potentially indicating optimal locations for future field studies.

  10. Compartmental model for organic matter digestion in facultative ponds.

    PubMed

    Giraldo, E; Garzón, A

    2002-01-01

    A model has been developed for the digestion of organic matter in facultative ponds in tropical regions. Complete mixing has been assumed for the aerobic and anaerobic compartments. Settling, aerobic layer oxidation, and anaerobic layer methanogenesis are the main processes for organic matter removal in the water column. Exchange processes between layers are dispersive or soluble exchange, solubilization and transport of organic matter from sediments to water column are also taken into account. Degradation of organic matter in the sediments produces gaseous emissions to the water column. The exchange between bubbles ascending and the water column was measured. The model was calibrated with data obtained from a pilot facultative pond built in Muña Reservoir in Bogotá. The pond was sampled during 4 months to compare data between its water hyacinth covered section and uncovered section. The results clearly show the relative importance of different BOD removal processes in facultative ponds and suggest modifications to further improve performance. The results from the model suggest that internal loadings to facultative ponds due to solubilization and return of organic matter from the sediments to the aerobic layer greatly influence the soluble BOD effluent concentration. Aerobic degradation activity in the facultative pond does not affect significantly the effluent concentration. Anaerobic degradation activity in the facultative pond can more easily achieve increases in the removal efficiencies of BOD.

  11. Compartmental model for organic matter digestion in facultative ponds.

    PubMed

    Giraldo, E; Garzón, A

    2002-01-01

    A model has been developed for the digestion of organic matter in facultative ponds in tropical regions. Complete mixing has been assumed for the aerobic and anaerobic compartments. Settling, aerobic layer oxidation, and anaerobic layer methanogenesis are the main processes for organic matter removal in the water column. Exchange processes between layers are dispersive or soluble exchange, solubilization and transport of organic matter from sediments to water column are also taken into account. Degradation of organic matter in the sediments produces gaseous emissions to the water column. The exchange between bubbles ascending and the water column was measured. The model was calibrated with data obtained from a pilot facultative pond built in Muña Reservoir in Bogotá. The pond was sampled during 4 months to compare data between its water hyacinth covered section and uncovered section. The results clearly show the relative importance of different BOD removal processes in facultative ponds and suggest modifications to further improve performance. The results from the model suggest that internal loadings to facultative ponds due to solubilization and return of organic matter from the sediments to the aerobic layer greatly influence the soluble BOD effluent concentration. Aerobic degradation activity in the facultative pond does not affect significantly the effluent concentration. Anaerobic degradation activity in the facultative pond can more easily achieve increases in the removal efficiencies of BOD. PMID:11833730

  12. Topsoil organic carbon content of Europe, a new map based on a generalised additive model

    NASA Astrophysics Data System (ADS)

    de Brogniez, Delphine; Ballabio, Cristiano; Stevens, Antoine; Jones, Robert J. A.; Montanarella, Luca; van Wesemael, Bas

    2014-05-01

    There is an increasing demand for up-to-date spatially continuous organic carbon (OC) data for global environment and climatic modeling. Whilst the current map of topsoil organic carbon content for Europe (Jones et al., 2005) was produced by applying expert-knowledge based pedo-transfer rules on large soil mapping units, the aim of this study was to replace it by applying digital soil mapping techniques on the first European harmonised geo-referenced topsoil (0-20 cm) database, which arises from the LUCAS (land use/cover area frame statistical survey) survey. A generalized additive model (GAM) was calibrated on 85% of the dataset (ca. 17 000 soil samples) and a backward stepwise approach selected slope, land cover, temperature, net primary productivity, latitude and longitude as environmental covariates (500 m resolution). The validation of the model (applied on 15% of the dataset), gave an R2 of 0.27. We observed that most organic soils were under-predicted by the model and that soils of Scandinavia were also poorly predicted. The model showed an RMSE of 42 g kg-1 for mineral soils and of 287 g kg-1 for organic soils. The map of predicted OC content showed the lowest values in Mediterranean countries and in croplands across Europe, whereas highest OC content were predicted in wetlands, woodlands and in mountainous areas. The map of standard error of the OC model predictions showed high values in northern latitudes, wetlands, moors and heathlands, whereas low uncertainty was mostly found in croplands. A comparison of our results with the map of Jones et al. (2005) showed a general agreement on the prediction of mineral soils' OC content, most probably because the models use some common covariates, namely land cover and temperature. Our model however failed to predict values of OC content greater than 200 g kg-1, which we explain by the imposed unimodal distribution of our model, whose mean is tilted towards the majority of soils, which are mineral. Finally, average

  13. Searching the MINT database for protein interaction information.

    PubMed

    Cesareni, Gianni; Chatr-aryamontri, Andrew; Licata, Luana; Ceol, Arnaud

    2008-06-01

    The Molecular Interactions Database (MINT) is a relational database designed to store information about protein interactions. Expert curators extract the relevant information from the scientific literature and deposit it in a computer readable form. Currently (April 2008), MINT contains information on almost 29,000 proteins and more than 100,000 interactions from more than 30 model organisms. This unit provides protocols for searching MINT over the Internet, using the MINT Viewer.

  14. There Is No Simple Model of the Plasma Membrane Organization

    PubMed Central

    Bernardino de la Serna, Jorge; Schütz, Gerhard J.; Eggeling, Christian; Cebecauer, Marek

    2016-01-01

    Ever since technologies enabled the characterization of eukaryotic plasma membranes, heterogeneities in the distributions of its constituents were observed. Over the years this led to the proposal of various models describing the plasma membrane organization such as lipid shells, picket-and-fences, lipid rafts, or protein islands, as addressed in numerous publications and reviews. Instead of emphasizing on one model we in this review give a brief overview over current models and highlight how current experimental work in one or the other way do not support the existence of a single overarching model. Instead, we highlight the vast variety of membrane properties and components, their influences and impacts. We believe that highlighting such controversial discoveries will stimulate unbiased research on plasma membrane organization and functionality, leading to a better understanding of this essential cellular structure. PMID:27747212

  15. Water quality modeling. (Latest citations from the NTIS database). Published Search

    SciTech Connect

    Not Available

    1993-03-01

    The bibliography contains citations concerning the application of mathematical modeling to hydrological and limnological systems. Nutrient removal in lakes and reservoirs, effects of mine drainage on water quality, and various parameters affecting pollutant flow in aquifers, streams, and rivers are discussed. Physical processes affecting water quality are included. (Contains a minimum of 180 citations and includes a subject term index and title list.)

  16. Water quality modeling. (Latest citations from the NTIS bibliographic database). Published Search

    SciTech Connect

    1995-01-01

    The bibliography contains citations concerning the application of mathematical modeling to hydrological and limnological systems. Nutrient removal in lakes and reservoirs, effects of mine drainage on water quality, and various parameters affecting pollutant flow in aquifers, streams, and rivers are discussed. Physical processes affecting water quality are included. (Contains a minimum of 173 citations and includes a subject term index and title list.)

  17. Development of a Conference Planning Model Using Integrated Database, Word Processing, and Spreadsheet Software.

    ERIC Educational Resources Information Center

    Stevens, William E.

    This report presents a model for conducting a statewide conference for the approximately 900 members of the South Carolina Council of Teachers of Mathematics (SCCTM) using the AppleWorks integrated software as the basis of the implementation plan. The first and second chapters provide background information on the conference and the…

  18. Green Algae as Model Organisms for Biological Fluid Dynamics*

    PubMed Central

    Goldstein, Raymond E.

    2015-01-01

    In the past decade the volvocine green algae, spanning from the unicellular Chlamydomonas to multicellular Volvox, have emerged as model organisms for a number of problems in biological fluid dynamics. These include flagellar propulsion, nutrient uptake by swimming organisms, hydrodynamic interactions mediated by walls, collective dynamics and transport within suspensions of microswimmers, the mechanism of phototaxis, and the stochastic dynamics of flagellar synchronization. Green algae are well suited to the study of such problems because of their range of sizes (from 10 μm to several millimetres), their geometric regularity, the ease with which they can be cultured and the availability of many mutants that allow for connections between molecular details and organism-level behavior. This review summarizes these recent developments and highlights promising future directions in the study of biological fluid dynamics, especially in the context of evolutionary biology, that can take advantage of these remarkable organisms. PMID:26594068

  19. Green Algae as Model Organisms for Biological Fluid Dynamics

    NASA Astrophysics Data System (ADS)

    Goldstein, Raymond E.

    2015-01-01

    In the past decade, the volvocine green algae, spanning from the unicellular Chlamydomonas to multicellular Volvox, have emerged as model organisms for a number of problems in biological fluid dynamics. These include flagellar propulsion, nutrient uptake by swimming organisms, hydrodynamic interactions mediated by walls, collective dynamics and transport within suspensions of microswimmers, the mechanism of phototaxis, and the stochastic dynamics of flagellar synchronization. Green algae are well suited to the study of such problems because of their range of sizes (from 10 μm to several millimeters), their geometric regularity, the ease with which they can be cultured, and the availability of many mutants that allow for connections between molecular details and organism-level behavior. This review summarizes these recent developments and highlights promising future directions in the study of biological fluid dynamics, especially in the context of evolutionary biology, that can take advantage of these remarkable organisms.

  20. Fast decision tree-based method to index large DNA-protein sequence databases using hybrid distributed-shared memory programming model.

    PubMed

    Jaber, Khalid Mohammad; Abdullah, Rosni; Rashid, Nur'Aini Abdul

    2014-01-01

    In recent times, the size of biological databases has increased significantly, with the continuous growth in the number of users and rate of queries; such that some databases have reached the terabyte size. There is therefore, the increasing need to access databases at the fastest rates possible. In this paper, the decision tree indexing model (PDTIM) was parallelised, using a hybrid of distributed and shared memory on resident database; with horizontal and vertical growth through Message Passing Interface (MPI) and POSIX Thread (PThread), to accelerate the index building time. The PDTIM was implemented using 1, 2, 4 and 5 processors on 1, 2, 3 and 4 threads respectively. The results show that the hybrid technique improved the speedup, compared to a sequential version. It could be concluded from results that the proposed PDTIM is appropriate for large data sets, in terms of index building time.

  1. An Updated One-Degree Seismic Tomographic Model Based on a Sensitivity Kernel Database

    NASA Astrophysics Data System (ADS)

    Sales de Andrade, E.; Liu, Q.; Ma, Z.; Manners, U.; Lee-Varisco, E.; Masters, G.

    2014-12-01

    Seismic tomography is instrumental in mapping 3D velocity structures of the Earth's interior based on travel-time measurements and waveform differences. Although both ray theory and other asymptotic methods have been successfully employed in global tomography, they are less accurate for long-period waves or steep velocity gradients. They also lack the ability to predict "non-geometrical" effects such as those for the core diffracted phases (Pdiff , Sdiff) which are crucial for mapping heterogeneities in the lowermost mantle (D'' layer). On the other hand, sensitivity kernels can be accurately calculated with no approximations by the interaction of forward and adjoint wavefields, both numerically simulated by spectral element methods. We have previously shown that by taking advantage of the symmetry of 1D reference models, we can efficiently and speedily construct sensitivity kernels of both P and S wavespeeds based on the simulation and storage of forward and adjoint strain fields for select source and receiver geometries. We have also shown the effect on global models inverted based on ray theory versus finite-frequency kernels in a parallel LSQR inversion. The CRUST1.0 model provides an updated version of the global sediment thickness on a finer 1° grid, based on active source seismic studies as well as receiver function studies. Crustal corrections for CRUST1.0 for the mantle may vary significantly from the ray-theoretical values, depending on crustal thickness, crustal velocities and wave period. We improve upon our previous model by incorporating the CRUST1.0 sediments and crustal corrections. This provides better response in the upper mantle and crustal regions. We produce this updated model for the S phases, and intend to extend the results to both P and S phases in the future.

  2. Modelling the fate of organic micropollutants in stormwater ponds.

    PubMed

    Vezzaro, Luca; Eriksson, Eva; Ledin, Anna; Mikkelsen, Peter S

    2011-06-01

    Urban water managers need to estimate the potential removal of organic micropollutants (MP) in stormwater treatment systems to support MP pollution control strategies. This study documents how the potential removal of organic MP in stormwater treatment systems can be quantified by using multimedia models. The fate of four different MP in a stormwater retention pond was simulated by applying two steady-state multimedia fate models (EPI Suite and SimpleBox) commonly applied in chemical risk assessment and a dynamic multimedia fate model (Stormwater Treatment Unit Model for Micro Pollutants--STUMP). The four simulated organic stormwater MP (iodopropynyl butylcarbamate--IPBC, benzene, glyphosate and pyrene) were selected according to their different urban sources and environmental fate. This ensures that the results can be extended to other relevant stormwater pollutants. All three models use substance inherent properties to calculate MP fate but differ in their ability to represent the small physical scale and high temporal variability of stormwater treatment systems. Therefore the three models generate different results. A Global Sensitivity Analysis (GSA) highlighted that settling/resuspension of particulate matter was the most sensitive process for the dynamic model. The uncertainty of the estimated MP fluxes can be reduced by calibrating the dynamic model against total suspended solids data. This reduction in uncertainty was more significant for the substances with strong tendency to sorb, i.e. glyphosate and pyrene and less significant for substances with a smaller tendency to sorb, i.e. IPBC and benzene. The results provide support to the elaboration of MP pollution control strategies by limiting the need for extensive and complex monitoring campaigns targeting the wide range of specific organic MP found in stormwater runoff. PMID:21496881

  3. Modelling the fate of organic micropollutants in stormwater ponds.

    PubMed

    Vezzaro, Luca; Eriksson, Eva; Ledin, Anna; Mikkelsen, Peter S

    2011-06-01

    Urban water managers need to estimate the potential removal of organic micropollutants (MP) in stormwater treatment systems to support MP pollution control strategies. This study documents how the potential removal of organic MP in stormwater treatment systems can be quantified by using multimedia models. The fate of four different MP in a stormwater retention pond was simulated by applying two steady-state multimedia fate models (EPI Suite and SimpleBox) commonly applied in chemical risk assessment and a dynamic multimedia fate model (Stormwater Treatment Unit Model for Micro Pollutants--STUMP). The four simulated organic stormwater MP (iodopropynyl butylcarbamate--IPBC, benzene, glyphosate and pyrene) were selected according to their different urban sources and environmental fate. This ensures that the results can be extended to other relevant stormwater pollutants. All three models use substance inherent properties to calculate MP fate but differ in their ability to represent the small physical scale and high temporal variability of stormwater treatment systems. Therefore the three models generate different results. A Global Sensitivity Analysis (GSA) highlighted that settling/resuspension of particulate matter was the most sensitive process for the dynamic model. The uncertainty of the estimated MP fluxes can be reduced by calibrating the dynamic model against total suspended solids data. This reduction in uncertainty was more significant for the substances with strong tendency to sorb, i.e. glyphosate and pyrene and less significant for substances with a smaller tendency to sorb, i.e. IPBC and benzene. The results provide support to the elaboration of MP pollution control strategies by limiting the need for extensive and complex monitoring campaigns targeting the wide range of specific organic MP found in stormwater runoff.

  4. Modeling organic matter stabilization during windrow composting of livestock effluents.

    PubMed

    Oudart, D; Paul, E; Robin, P; Paillat, J M

    2012-01-01

    Composting is a complex bioprocess, requiring a lot of empirical experiments to optimize the process. A dynamical mathematical model for the biodegradation of the organic matter during the composting process has been developed. The initial organic matter expressed by chemical oxygen demand (COD) is decomposed into rapidly and slowly degraded compartments and an inert one. The biodegradable COD is hydrolysed and consumed by microorganisms and produces metabolic water and carbon dioxide. This model links a biochemical characterization of the organic matter by Van Soest fractionating with COD. The comparison of experimental and simulation results for carbon dioxide emission, dry matter and carbon content balance showed good correlation. The initial sizes of the biodegradable COD compartments are explained by the soluble, hemicellulose-like and lignin fraction. Their sizes influence the amplitude of the carbon dioxide emission peak. The initial biomass is a sensitive variable too, influencing the time at which the emission peak occurs. PMID:23393964

  5. A Dynamical Phyllotaxis Model to Determine Floral Organ Number

    PubMed Central

    Kitazawa, Miho S.; Fujimoto, Koichi

    2015-01-01

    How organisms determine particular organ numbers is a fundamental key to the development of precise body structures; however, the developmental mechanisms underlying organ-number determination are unclear. In many eudicot plants, the primordia of sepals and petals (the floral organs) first arise sequentially at the edge of a circular, undifferentiated region called the floral meristem, and later transition into a concentric arrangement called a whorl, which includes four or five organs. The properties controlling the transition to whorls comprising particular numbers of organs is little explored. We propose a development-based model of floral organ-number determination, improving upon earlier models of plant phyllotaxis that assumed two developmental processes: the sequential initiation of primordia in the least crowded space around the meristem and the constant growth of the tip of the stem. By introducing mutual repulsion among primordia into the growth process, we numerically and analytically show that the whorled arrangement emerges spontaneously from the sequential initiation of primordia. Moreover, by allowing the strength of the inhibition exerted by each primordium to decrease as the primordium ages, we show that pentamerous whorls, in which the angular and radial positions of the primordia are consistent with those observed in sepal and petal primordia in Silene coeli-rosa, Caryophyllaceae, become the dominant arrangement. The organ number within the outmost whorl, corresponding to the sepals, takes a value of four or five in a much wider parameter space than that in which it takes a value of six or seven. These results suggest that mutual repulsion among primordia during growth and a temporal decrease in the strength of the inhibition during initiation are required for the development of the tetramerous and pentamerous whorls common in eudicots. PMID:25950739

  6. The Halophile protein database.

    PubMed

    Sharma, Naveen; Farooqi, Mohammad Samir; Chaturvedi, Krishna Kumar; Lal, Shashi Bhushan; Grover, Monendra; Rai, Anil; Pandey, Pankaj

    2014-01-01

    Halophilic archaea/bacteria adapt to different salt concentration, namely extreme, moderate and low. These type of adaptations may occur as a result of modification of protein structure and other changes in different cell organelles. Thus proteins may play an important role in the adaptation of halophilic archaea/bacteria to saline conditions. The Halophile protein database (HProtDB) is a systematic attempt to document the biochemical and biophysical properties of proteins from halophilic archaea/bacteria which may be involved in adaptation of these organisms to saline conditions. In this database, various physicochemical properties such as molecular weight, theoretical pI, amino acid composition, atomic composition, estimated half-life, instability index, aliphatic index and grand average of hydropathicity (Gravy) have been listed. These physicochemical properties play an important role in identifying the protein structure, bonding pattern and function of the specific proteins. This database is comprehensive, manually curated, non-redundant catalogue of proteins. The database currently contains 59 897 proteins properties extracted from 21 different strains of halophilic archaea/bacteria. The database can be accessed through link. Database URL: http://webapp.cabgrid.res.in/protein/

  7. Screening and analysis of 940 organic micro-pollutants in river sediments in Vietnam using an automated identification and quantification database system for GC-MS.

    PubMed

    Duong, Hanh Thi; Kadokami, Kiwao; Pan, Shuangye; Matsuura, Naoki; Nguyen, Trung Quang

    2014-07-01

    In order to obtain a detailed picture of pollution by organic micro-pollutants in Vietnamese rivers, 940 semi-volatile organic compounds in river sediments collected from four major cities were examined by a comprehensive gas chromatography-mass spectrometry-database. The number of detected chemicals at each site ranged from 49 to 158 (median 96 out of 940) with 185 analytes detected at least once in the survey. The substances detected with high frequency (over 80%) and high concentrations were n-alkanes, phthalates, sterols and PAHs. For most substances, sediments from metropolitan areas (Hanoi and Ho Chi Minh City) were more heavily contaminated than those in rural and suburban areas. Sterols were observed in nearly 100% of sediments at extremely high concentrations, suggesting that the studied rivers were contaminated by sewage. Pyrethroids (permethrin-1 and -2) were the most dominant insecticides found in inner canals of Hanoi and Ho Chi Minh City. Deltamethrin was only detected at a site in Hanoi at an elevated concentration. This reflects that pyrethroids are used for the protection of private and public health rather than for agriculture. p,p'-DDE and p,p'-DDD were the dominant members of the DDT family of chemicals detected, indicating no recent inputs of DDTs in the study areas. PCBs residues were lower than those in other Asian countries, which suggest historically much lower use of PCBs in Vietnam. PAHs pollution in urban areas is caused by the runoff of petroleum products and vehicle exhaust gases, whereas in rural and suburban areas, the combustion of fossil fuels and biomass is major sources of PAHs. Overall, the study confirmed that rivers in Vietnam were heavily polluted mainly by domestic wastewater.

  8. Comprehensive mollusk acute toxicity database improves the use of Interspecies Correlation Estimation (ICE) models to predict toxicity of untested freshwater and endangered mussel species

    EPA Science Inventory

    Interspecies correlation estimation (ICE) models extrapolate acute toxicity data from surrogate test species to untested taxa. A suite of ICE models developed from a comprehensive database is available on the US Environmental Protection Agency’s web-based application, Web-I...

  9. A Process Model for the Comprehension of Organic Chemistry Notation

    ERIC Educational Resources Information Center

    Havanki, Katherine L.

    2012-01-01

    This dissertation examines the cognitive processes individuals use when reading organic chemistry equations and factors that affect these processes, namely, visual complexity of chemical equations and participant characteristics (expertise, spatial ability, and working memory capacity). A six stage process model for the comprehension of organic…

  10. Promoting Representational Competence with Molecular Models in Organic Chemistry

    ERIC Educational Resources Information Center

    Stull, Andrew T.; Gainer, Morgan; Padalkar, Shamin; Hegarty, Mary

    2016-01-01

    Mastering the many different diagrammatic representations of molecules used in organic chemistry is challenging for students. This article summarizes recent research showing that manipulating 3-D molecular models can facilitate the understanding and use of these representations. Results indicate that students are more successful in translating…

  11. An Integrated Model for Effective Knowledge Management in Chinese Organizations

    ERIC Educational Resources Information Center

    An, Xiaomi; Deng, Hepu; Wang, Yiwen; Chao, Lemen

    2013-01-01

    Purpose: The purpose of this paper is to provide organizations in the Chinese cultural context with a conceptual model for an integrated adoption of existing knowledge management (KM) methods and to improve the effectiveness of their KM activities. Design/methodology/approaches: A comparative analysis is conducted between China and the western…

  12. Modeling emissions of volatile organic compounds from silage

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Volatile organic compounds (VOCs), necessary reactants for photochemical smog formation, are emitted from numerous sources. Limited available data suggest that dairy farms emit VOCs with cattle feed, primarily silage, being the primary source. Process-based models of VOC transfer within and from si...

  13. A data-based model to locate mass movements triggered by seismic events in Sichuan, China.

    PubMed

    de Souza, Fabio Teodoro

    2014-01-01

    Earthquakes affect the entire world and have catastrophic consequences. On May 12, 2008, an earthquake of magnitude 7.9 on the Richter scale occurred in the Wenchuan area of Sichuan province in China. This event, together with subsequent aftershocks, caused many avalanches, landslides, debris flows, collapses, and quake lakes and induced numerous unstable slopes. This work proposes a methodology that uses a data mining approach and geographic information systems to predict these mass movements based on their association with the main and aftershock epicenters, geologic faults, riverbeds, and topography. A dataset comprising 3,883 mass movements is analyzed, and some models to predict the location of these mass movements are developed. These predictive models could be used by the Chinese authorities as an important tool for identifying risk areas and rescuing survivors during similar events in the future.

  14. A data-based model to locate mass movements triggered by seismic events in Sichuan, China.

    PubMed

    de Souza, Fabio Teodoro

    2014-01-01

    Earthquakes affect the entire world and have catastrophic consequences. On May 12, 2008, an earthquake of magnitude 7.9 on the Richter scale occurred in the Wenchuan area of Sichuan province in China. This event, together with subsequent aftershocks, caused many avalanches, landslides, debris flows, collapses, and quake lakes and induced numerous unstable slopes. This work proposes a methodology that uses a data mining approach and geographic information systems to predict these mass movements based on their association with the main and aftershock epicenters, geologic faults, riverbeds, and topography. A dataset comprising 3,883 mass movements is analyzed, and some models to predict the location of these mass movements are developed. These predictive models could be used by the Chinese authorities as an important tool for identifying risk areas and rescuing survivors during similar events in the future. PMID:24085622

  15. Water quality modeling. (Latest citations from the NTIS Bibliographic database). Published Search

    SciTech Connect

    1996-01-01

    The bibliography contains citations concerning the application of mathematical modeling to hydrological and limnological systems. Nutrient removal in lakes and reservoirs, effects of mine drainage on water quality, and various parameters affecting pollutant flow in aquifers, streams, and rivers are discussed. Physical processes affecting water quality are included. (Contains 50-250 citations and includes a subject term index and title list.) (Copyright NERAC, Inc. 1995)

  16. Model checking software for phylogenetic trees using distribution and database methods.

    PubMed

    Requeno, José Ignacio; Colom, José Manuel

    2013-01-01

    Model checking, a generic and formal paradigm stemming from computer science based on temporal logics, has been proposed for the study of biological properties that emerge from the labeling of the states defined over the phylogenetic tree. This strategy allows us to use generic software tools already present in the industry. However, the performance of traditional model checking is penalized when scaling the system for large phylogenies. To this end, two strategies are presented here. The first one consists of partitioning the phylogenetic tree into a set of subgraphs each one representing a subproblem to be verified so as to speed up the computation time and distribute the memory consumption. The second strategy is based on uncoupling the information associated to each state of the phylogenetic tree (mainly, the DNA sequence) and exporting it to an external tool for the management of large information systems. The integration of all these approaches outperforms the results of monolithic model checking and helps us to execute the verification of properties in a real phylogenetic tree. PMID:24231143

  17. A general framework for modelling the vertical organic matter profile in mineral and organic soils

    NASA Astrophysics Data System (ADS)

    Braakhekke, Maarten; Ahrens, Bernhard

    2016-04-01

    The vertical distribution of soil organic matter (SOM) within the mineral soil and surface organic layer is an important property of terrestrial ecosystems that affects carbon and nutrient cycling and soil heat and moisture transport. The overwhelming majority of models of SOM dynamics are zero-dimensional, i.e. they do not resolve heterogeneity of SOM concentration along the vertical profile. In recent years, however, a number of new vertically explicit SOM models or vertically explicit versions of existing models have been published. These models describe SOM in units of concentration (mass per unit volume) by means of a reactive-transport model that includes diffusion and/or advection terms for SOM transport, and vertically resolves SOM inputs and factors that influence decomposition. An important assumption behind these models is that the volume of soil elements is constant over time, i.e. not affected by SOM dynamics. This assumption only holds if the SOM content is negligible compared to the mineral content. When this is not the case, SOM input or loss in a soil element may cause a change in volume of the element rather than a change in SOM concentration. Furthermore, these volume changes can cause vertical shifts of material relative to the surface. This generally causes material in an organic layer to gradually move downward, even in absence of mixing processes. Since the classical reactive-transport model of the SOM profile can only be applied to the mineral soil, the surface organic layer is usually either treated separately or not explicitly considered. We present a new and elegant framework that treats the surface organic layer and mineral soil as one continuous whole. It explicitly accounts for volume changes due to SOM dynamics and changes in bulk density. The vertical shifts resulting from these volume changes are included in an Eulerian representation as an additional advective transport flux. Our approach offers a more elegant and realistic

  18. Lamination of organic solar cells and organic light emitting devices: Models and experiments

    SciTech Connect

    Oyewole, O. K.; Yu, D.; Du, J.; Asare, J.; Fashina, A.; Anye, V. C.; Zebaze Kana, M. G.; Soboyejo, W. O.

    2015-08-21

    In this paper, a combined experimental, computational, and analytical approach is used to provide new insights into the lamination of organic solar cells and light emitting devices at macro- and micro-scales. First, the effects of applied lamination force (on contact between the laminated layers) are studied. The crack driving forces associated with the interfacial cracks (at the bi-material interfaces) are estimated along with the critical interfacial crack driving forces associated with the separation of thin films, after layer transfer. The conditions for successful lamination are predicted using a combination of experiments and computational models. Guidelines are developed for the lamination of low-cost organic electronic structures.

  19. Unexpected capacity for organic carbon assimilation by Thermosynechococcus elongatus, a crucial photosynthetic model organism.

    PubMed

    Zilliges, Yvonne; Dau, Holger

    2016-04-01

    Genetic modification of key residues of photosystems is essential to identify functionally crucial processes by spectroscopic and crystallographic investigation; the required protein stability favours use of thermophilic species. The currently unique thermophilic photosynthetic model organism is the cyanobacterial genus Thermosynechococcus. We report the ability of Thermosynechococcus elongatus to assimilate organic carbon, specifically D-fructose. Growth in the presence of a photosynthesis inhibitor opens the door towards crucial amino acid substitutions in photosystems by the rescue of otherwise lethal mutations. Yet depression of batch-culture growth after 7 days implies that additional developments are needed. PMID:26935247

  20. Modeling 4D Changes in Pathological Anatomy using Domain Adaptation: Analysis of TBI Imaging using a Tumor Database.

    PubMed

    Wang, Bo; Prastawa, Marcel; Saha, Avishek; Awate, Suyash P; Irimia, Andrei; Chambers, Micah C; Vespa, Paul M; Van Horn, John D; Pascucci, Valerio; Gerig, Guido

    2013-01-01

    Analysis of 4D medical images presenting pathology (i.e., lesions) is significantly challenging due to the presence of complex changes over time. Image analysis methods for 4D images with lesions need to account for changes in brain structures due to deformation, as well as the formation and deletion of new structures (e.g., edema, bleeding) due to the physiological processes associated with damage, intervention, and recovery. We propose a novel framework that models 4D changes in pathological anatomy across time, and provides explicit mapping from a healthy template to subjects with pathology. Moreover, our framework uses transfer learning to leverage rich information from a known source domain, where we have a collection of completely segmented images, to yield effective appearance models for the input target domain. The automatic 4D segmentation method uses a novel domain adaptation technique for generative kernel density models to transfer information between different domains, resulting in a fully automatic method that requires no user interaction. We demonstrate the effectiveness of our novel approach with the analysis of 4D images of traumatic brain injury (TBI), using a synthetic tumor database as the source domain. PMID:25346953