Sample records for repository research experiments

  1. Evolution of a Digital Repository: One Institution's Experience

    ERIC Educational Resources Information Center

    Owen, Terry M.

    2011-01-01

    In this article, the development of a digital repository is examined, specifically how the focus on acquiring content for the repository has transitioned from faculty-published research to include the gray literature produced by the research centers on campus, including unpublished technical reports and undergraduate research from honors programs.…

  2. Research Students and the Loughborough Institutional Repository

    ERIC Educational Resources Information Center

    Pickton, Margaret; McKnight, Cliff

    2006-01-01

    This article investigates the potential role for research students in an institutional repository (IR). Face-to-face interviews with 34 research students at Loughborough University were carried out. Using a mixture of closed and open questions, the interviews explored the students' experiences and opinions of publishing, open access and the…

  3. Communicating the Value of an Institutional Repository: Experiences at Ghana's University for Development Studies

    ERIC Educational Resources Information Center

    Thompson, Edwin S.; Akeriwe, Miriam Linda; Aikins, Angela Achia

    2016-01-01

    The quality of research depends greatly on access to existing information. Institutional repositories (IRs) have the potential to enhance and promote the dissemination of knowledge and research. This may lead to discoveries and innovation alongside maximizing return on investment in research and development. Following some background information,…

  4. A Comparison of Subject and Institutional Repositories in Self-Archiving Practices

    ERIC Educational Resources Information Center

    Xia, Jingfeng

    2008-01-01

    The disciplinary culture theory presumes that if a scholar has been familiar with self-archiving through an existing subject-based repository, this scholar will be more enthusiastic about contributing his/her research to an institutional repository than one who has not had the experience. To test the theory, this article examines self-archiving…

  5. The Use of Underground Research Laboratories to Support Repository Development Programs. A Roadmap for the Underground Research Facilities Network.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    MacKinnon, Robert J.

    2015-10-26

    Under the auspices of the International Atomic Energy Agency (IAEA), nationally developed underground research laboratories (URLs) and associated research institutions are being offered for use by other nations. These facilities form an Underground Research Facilities (URF) Network for training in and demonstration of waste disposal technologies and the sharing of knowledge and experience related to geologic repository development, research, and engineering. In order to achieve its objectives, the URF Network regularly sponsors workshops and training events related to the knowledge base that is transferable between existing URL programs and to nations with an interest in developing a new URL. Thismore » report describes the role of URLs in the context of a general timeline for repository development. This description includes identification of key phases and activities that contribute to repository development as a repository program evolves from an early research and development phase to later phases such as construction, operations, and closure. This information is cast in the form of a matrix with the entries in this matrix forming the basis of the URF Network roadmap that will be used to identify and plan future workshops and training events.« less

  6. Mont Terri Underground Rock Laboratory, Switzerland-Research Program And Key Results

    NASA Astrophysics Data System (ADS)

    Nussbaum, C. O.; Bossart, P. J.

    2012-12-01

    Argillaceous formations generally act as aquitards because of their low hydraulic conductivities. This property, together with the large retention capacity of clays for cationic contaminants and the potential for self-sealing, has brought clay formations into focus as potential host rocks for the geological disposal of radioactive waste. Excavated in the Opalinus Clay formation, the Mont Terri underground rock laboratory in the Jura Mountains of NW Switzerland is an important international test site for researching clay formations. Research is carried out in the underground facility, which is located adjacent to the security gallery of the Mont Terri motorway tunnel. Fifteen partners from European countries, USA, Canada and Japan participate in the project. The objectives of the research program are to analyze the hydrogeological, geochemical and rock mechanical properties of the Opalinus Clay, to determine the changes induced by the excavation of galleries and by heating of the rock formation, to test sealing and container emplacement techniques and to evaluate and improve suitable investigation techniques. For the safety of deep geological disposal, it is of key importance to understand the processes occurring in the undisturbed argillaceous environment, as well as the processes in a disturbed system, during the operation of the repository. The objectives are related to: 1. Understanding processes and mechanisms in undisturbed clays and 2. Experiments related to repository-induced perturbations. Experiments of the first group are dedicated to: i) Improvement of drilling and excavation technologies and sampling methods; ii) Estimation of hydrogeological, rock mechanical and geochemical parameters of the undisturbed Opalinus Clay. Upscaling of parameters from laboratory to in situ scale; iii) Geochemistry of porewater and natural gases; evolution of porewater over time scales; iv) Assessment of long-term hydraulic transients associated with erosion and thermal scenarios and v) Evaluation of diffusion and retention parameters for long-lived radionuclides. Experiments related to repository-induced perturbations are focused on: i) Influence of rock liner on the disposal system and the buffering potential of the host rock; ii) Self-sealing processes in the excavation damaged zone; iii) Hydro-mechanical coupled processes (e.g. stress redistributions and pore pressure evolution during excavation); iv) Thermo-hydro-mechanical-chemical coupled processes (e.g. heating of bentonite and host rock) and v) Gas-induced transport of radionuclides in porewater and along interfaces in the engineered barrier system. A third research direction is to demonstrate the feasibility of repository construction and long-term safety after repository closure. Demonstration experiments can contribute to improving the reliability of the scientific basis for the safety assessment of future geological repositories, particularly if they are performed on a large scale and with a long duration. These experiments include the construction and installation of engineered barriers on a 1:1 scale: i) Horizontal emplacement of canisters; ii) Evaluation of the corrosion of container materials; repository re-saturation; iii) Sealing of boreholes and repository access tunnels and iv) Long-term monitoring of the repository. References Bossart, P. & Thury, M. (2008): Mont Terri Rock Laboratory. Project, Programme 1996 to 2007 and Results. - Rep. Swiss Geol. Surv. 3.

  7. jPOSTrepo: an international standard data repository for proteomes

    PubMed Central

    Okuda, Shujiro; Watanabe, Yu; Moriya, Yuki; Kawano, Shin; Yamamoto, Tadashi; Matsumoto, Masaki; Takami, Tomoyo; Kobayashi, Daiki; Araki, Norie; Yoshizawa, Akiyasu C.; Tabata, Tsuyoshi; Sugiyama, Naoyuki; Goto, Susumu; Ishihama, Yasushi

    2017-01-01

    Major advancements have recently been made in mass spectrometry-based proteomics, yielding an increasing number of datasets from various proteomics projects worldwide. In order to facilitate the sharing and reuse of promising datasets, it is important to construct appropriate, high-quality public data repositories. jPOSTrepo (https://repository.jpostdb.org/) has successfully implemented several unique features, including high-speed file uploading, flexible file management and easy-to-use interfaces. This repository has been launched as a public repository containing various proteomic datasets and is available for researchers worldwide. In addition, our repository has joined the ProteomeXchange consortium, which includes the most popular public repositories such as PRIDE in Europe for MS/MS datasets and PASSEL for SRM datasets in the USA. Later MassIVE was introduced in the USA and accepted into the ProteomeXchange, as was our repository in July 2016, providing important datasets from Asia/Oceania. Accordingly, this repository thus contributes to a global alliance to share and store all datasets from a wide variety of proteomics experiments. Thus, the repository is expected to become a major repository, particularly for data collected in the Asia/Oceania region. PMID:27899654

  8. ORIOLE, in the Search for Evidence of OER in Teaching. Experiences in the Use, Re-Use and the Sharing and Influence of Repositories

    ERIC Educational Resources Information Center

    Santos-Hermosa, Gema

    2014-01-01

    The study presented here aims to gather useful information on the use, re-reuse and sharing of resources in Education and also the influence of repositories, to better understand the perspective of individual practitioners and suggest future areas of debate for researchers. Open Resources: Influence on Learners and Educators (ORIOLE) project, was…

  9. Making research data repositories visible: the re3data.org Registry.

    PubMed

    Pampel, Heinz; Vierkant, Paul; Scholze, Frank; Bertelmann, Roland; Kindling, Maxi; Klump, Jens; Goebelbecker, Hans-Jürgen; Gundlach, Jens; Schirmbacher, Peter; Dierolf, Uwe

    2013-01-01

    Researchers require infrastructures that ensure a maximum of accessibility, stability and reliability to facilitate working with and sharing of research data. Such infrastructures are being increasingly summarized under the term Research Data Repositories (RDR). The project re3data.org-Registry of Research Data Repositories-has begun to index research data repositories in 2012 and offers researchers, funding organizations, libraries and publishers an overview of the heterogeneous research data repository landscape. In July 2013 re3data.org lists 400 research data repositories and counting. 288 of these are described in detail using the re3data.org vocabulary. Information icons help researchers to easily identify an adequate repository for the storage and reuse of their data. This article describes the heterogeneous RDR landscape and presents a typology of institutional, disciplinary, multidisciplinary and project-specific RDR. Further the article outlines the features of re3data.org, and shows how this registry helps to identify appropriate repositories for storage and search of research data.

  10. myExperiment: a repository and social network for the sharing of bioinformatics workflows

    PubMed Central

    Goble, Carole A.; Bhagat, Jiten; Aleksejevs, Sergejs; Cruickshank, Don; Michaelides, Danius; Newman, David; Borkum, Mark; Bechhofer, Sean; Roos, Marco; Li, Peter; De Roure, David

    2010-01-01

    myExperiment (http://www.myexperiment.org) is an online research environment that supports the social sharing of bioinformatics workflows. These workflows are procedures consisting of a series of computational tasks using web services, which may be performed on data from its retrieval, integration and analysis, to the visualization of the results. As a public repository of workflows, myExperiment allows anybody to discover those that are relevant to their research, which can then be reused and repurposed to their specific requirements. Conversely, developers can submit their workflows to myExperiment and enable them to be shared in a secure manner. Since its release in 2007, myExperiment currently has over 3500 registered users and contains more than 1000 workflows. The social aspect to the sharing of these workflows is facilitated by registered users forming virtual communities bound together by a common interest or research project. Contributors of workflows can build their reputation within these communities by receiving feedback and credit from individuals who reuse their work. Further documentation about myExperiment including its REST web service is available from http://wiki.myexperiment.org. Feedback and requests for support can be sent to bugs@myexperiment.org. PMID:20501605

  11. Evaluation Methodologies for Information Management Systems; Building Digital Tobacco Industry Document Libraries at the University of California, San Francisco Library/Center for Knowledge Management; Experiments with the IFLA Functional Requirements for Bibliographic Records (FRBR); Coming to Term: Designing the Texas Email Repository Model.

    ERIC Educational Resources Information Center

    Morse, Emile L.; Schmidt, Heidi; Butter, Karen; Rider, Cynthia; Hickey, Thomas B.; O'Neill, Edward T.; Toves, Jenny; Green, Marlan; Soy, Sue; Gunn, Stan; Galloway, Patricia

    2002-01-01

    Includes four articles that discuss evaluation methods for information management systems under the Defense Advanced Research Projects Agency; building digital libraries at the University of California San Francisco's Tobacco Control Archives; IFLA's Functional Requirements for Bibliographic Records; and designing the Texas email repository model…

  12. Terminology development towards harmonizing multiple clinical neuroimaging research repositories.

    PubMed

    Turner, Jessica A; Pasquerello, Danielle; Turner, Matthew D; Keator, David B; Alpert, Kathryn; King, Margaret; Landis, Drew; Calhoun, Vince D; Potkin, Steven G; Tallis, Marcelo; Ambite, Jose Luis; Wang, Lei

    2015-07-01

    Data sharing and mediation across disparate neuroimaging repositories requires extensive effort to ensure that the different domains of data types are referred to by commonly agreed upon terms. Within the SchizConnect project, which enables querying across decentralized databases of neuroimaging, clinical, and cognitive data from various studies of schizophrenia, we developed a model for each data domain, identified common usable terms that could be agreed upon across the repositories, and linked them to standard ontological terms where possible. We had the goal of facilitating both the current user experience in querying and future automated computations and reasoning regarding the data. We found that existing terminologies are incomplete for these purposes, even with the history of neuroimaging data sharing in the field; and we provide a model for efforts focused on querying multiple clinical neuroimaging repositories.

  13. Terminology development towards harmonizing multiple clinical neuroimaging research repositories

    PubMed Central

    Turner, Jessica A.; Pasquerello, Danielle; Turner, Matthew D.; Keator, David B.; Alpert, Kathryn; King, Margaret; Landis, Drew; Calhoun, Vince D.; Potkin, Steven G.; Tallis, Marcelo; Ambite, Jose Luis; Wang, Lei

    2015-01-01

    Data sharing and mediation across disparate neuroimaging repositories requires extensive effort to ensure that the different domains of data types are referred to by commonly agreed upon terms. Within the SchizConnect project, which enables querying across decentralized databases of neuroimaging, clinical, and cognitive data from various studies of schizophrenia, we developed a model for each data domain, identified common usable terms that could be agreed upon across the repositories, and linked them to standard ontological terms where possible. We had the goal of facilitating both the current user experience in querying and future automated computations and reasoning regarding the data. We found that existing terminologies are incomplete for these purposes, even with the history of neuroimaging data sharing in the field; and we provide a model for efforts focused on querying multiple clinical neuroimaging repositories. PMID:26688838

  14. Making Research Data Repositories Visible: The re3data.org Registry

    PubMed Central

    Pampel, Heinz; Vierkant, Paul; Scholze, Frank; Bertelmann, Roland; Kindling, Maxi; Klump, Jens; Goebelbecker, Hans-Jürgen; Gundlach, Jens; Schirmbacher, Peter; Dierolf, Uwe

    2013-01-01

    Researchers require infrastructures that ensure a maximum of accessibility, stability and reliability to facilitate working with and sharing of research data. Such infrastructures are being increasingly summarized under the term Research Data Repositories (RDR). The project re3data.org–Registry of Research Data Repositories–has begun to index research data repositories in 2012 and offers researchers, funding organizations, libraries and publishers an overview of the heterogeneous research data repository landscape. In July 2013 re3data.org lists 400 research data repositories and counting. 288 of these are described in detail using the re3data.org vocabulary. Information icons help researchers to easily identify an adequate repository for the storage and reuse of their data. This article describes the heterogeneous RDR landscape and presents a typology of institutional, disciplinary, multidisciplinary and project-specific RDR. Further the article outlines the features of re3data.org, and shows how this registry helps to identify appropriate repositories for storage and search of research data. PMID:24223762

  15. Data-intensive science gateway for rock physicists and volcanologists.

    NASA Astrophysics Data System (ADS)

    Filgueira, Rosa; Atkinson, Malcom; Bell, Andrew; Main, Ian; Boon, Steve; Meredith, Philp; Kilburn, Christopher

    2014-05-01

    Scientists have always shared data and mathematical models of the phenomena they study. Rock physics and Volcanology, as well as other solid-Earth sciences, have increasingly used Internet communications and computational renditions of their models for this purpose over the last two decades. Here we consider how to organise rock physics and volcanology data to open up opportunities for sharing and comparing both experiment data from experiments, observations and model runs and analytic interpretations of these data. Our hypothesis is that if we facilitate productive information sharing across those communities by using a new science gateway, it will benefit the science. The proposed science gateway should make the first steps for making existing research practices easier and facilitate new research. It will achieve this by supporting three major functions: 1) sharing data from laboratories and observatories, experimental facilities and models; 2) sharing models of rock fracture and methods for analysing experimental data; and 3) supporting recurrent operational tasks, such as data collection and model application in real time. We report initial work in two projects (NERC EFFORT and NERC CREEP-2) and experience with an early web-accessible protytpe called EFFORT gateway, where we are implementing such information sharing services for those projects. 1. Sharing data: In EFFORT gateway, we are working on several facilities for sharing data: *Upload data: We have designed and developed a new adaptive data transfer java tool called FAST (Flexible Automated Streaming Transfer) to upload experimental data and metadata periodically from laboratories to our repository. *Visualisation: As data are deposited in the repository, a visualisation of the accumulated data is made available for display in the Web portal. *Metadata and catalogues: The gateway uses a repository to hold all the data and a catalogue to hold all the corresponding metadata. 2. Sharing models and methods: The EFFORT gateway uses a repository to hold all of the models and a catalogue to hold the corresponding metadata. It provides several Web facilities for uploading, accessing and testing models. *Upload and store models: Through the gateway, researchers can upload as many models to the repository as they want. *Description of models: The gateway solicits and creates metadata for every model uploaded to store in the catalogue. *Search for models: Researchers can search the catalogue for models by using prepackaged sql-queries. *Access to models: Once a researcher has selected the model(s) that is going to be used for analysing an experiment, it will be obtained from the gateway. *Services to test and run models: Once a researcher selects a model and the experimental data to which it should be applied, the gateway submits the corresponding computational job to a high-performance computational (HPC) resource hiding technical details. Once a job is submitted to the HPC cluster, the results are displayed in the gateway in real time, catalogued and stored in the data repository, allowing further researcher-instigated operations to retrieve, inspect and aggregate results. *Services to write models: We have desgined VarPy library, which is an open-source toolbox which provides a Python framework for analysing volcanology and rock physics data. It provides several functions, which allow users to define their own workflows to develop models, analyses and visualizations. 3. Recurrent Operations: We have started to introduce some recurrent operations: *Automated data upload: FAST provides a mechanism to automate the data upload. *Periodic activation of models: The EFFORT gateway allows researchers to run different models periodically against the experimental data that are being or have been uploaded

  16. Integration and Cooperation in the Next Golden Age of Human Space Flight Data Repositories: Tools for Retrospective Analysis and Future Planning

    NASA Technical Reports Server (NTRS)

    Thomas, D.; Fitts, M.; Wear, M.; VanBaalen, M.

    2011-01-01

    As NASA transitions from the Space Shuttle era into the next phase of space exploration, the need to ensure the capture, analysis, and application of its research and medical data is of greater urgency than at any other previous time. In this era of limited resources and challenging schedules, the Human Research Program (HRP) based at NASA s Johnson Space Center (JSC) recognizes the need to extract the greatest possible amount of information from the data already captured, as well as focus current and future research funding on addressing the HRP goal to provide human health and performance countermeasures, knowledge, technologies, and tools to enable safe, reliable, and productive human space exploration. To this end, the Science Management Office and the Medical Informatics and Health Care Systems Branch within the HRP and the Space Medicine Division have been working to make both research data and clinical data more accessible to the user community. The Life Sciences Data Archive (LSDA), the research repository housing data and information regarding the physiologic effects of microgravity, and the Lifetime Surveillance of Astronaut Health Repository (LSAH-R), the clinical repository housing astronaut data, have joined forces to achieve this goal. The task of both repositories is to acquire, preserve, and distribute data and information both within the NASA community and to the science community at large. This is accomplished via the LSDA s public website (http://lsda.jsc.nasa.gov), which allows access to experiment descriptions including hardware, datasets, key personnel, mission descriptions and a mechanism for researchers to request additional data, research and clinical, that is not accessible from the public website. This will result in making the work of NASA and its partners available to the wider sciences community, both domestic and international. The desired outcome is the use of these data for knowledge discovery, retrospective analysis, and planning of future research studies.

  17. The Amistad Research Center: Documenting the African American Experience.

    ERIC Educational Resources Information Center

    Chepesiuk, Ron

    1993-01-01

    Describes the Amistad Research Center housed at Tulane University which is a repository of primary documents on African-American history. Topics addressed include the development and growth of the collection; inclusion of the American Missionary Association archives; sources of support; civil rights; and collecting for the future. (LRW)

  18. Persistent Identifiers for Field Expeditions: A Next Step for the US Oceanographic Research Fleet

    NASA Astrophysics Data System (ADS)

    Arko, Robert; Carbotte, Suzanne; Chandler, Cynthia; Smith, Shawn; Stocks, Karen

    2016-04-01

    Oceanographic research cruises are complex affairs, typically requiring an extensive effort to secure the funding, plan the experiment, and mobilize the field party. Yet cruises are not typically published online as first-class digital objects with persistent, citable identifiers linked to the scientific literature. The Rolling Deck to Repository (R2R; info@rvdata.us) program maintains a master catalog of oceanographic cruises for the United States research fleet, currently documenting over 6,000 expeditions on 37 active and retired vessels. In 2015, R2R started routinely publishing a Digital Object Identifier (DOI) for each completed cruise. Cruise DOIs, in turn, are linked to related persistent identifiers where available including the Open Researcher and Contributor ID (ORCID) for members of the science party, the International Geo Sample Number (IGSN) for physical specimens collected during the cruise, the Open Funder Registry (FundRef) codes that supported the experiment, and additional DOIs for datasets, journal articles, and other products resulting from the cruise. Publishing a persistent identifier for each field expedition will facilitate interoperability between the many different repositories that hold research products from cruises; will provide credit to the investigators who secured the funding and carried out the experiment; and will facilitate the gathering of fleet-wide altmetrics that demonstrate the broad impact of oceanographic research.

  19. 10 CFR 60.44 - Changes, tests, and experiments.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... REPOSITORIES Licenses License Issuance and Amendment § 60.44 Changes, tests, and experiments. (a)(1) Following authorization to receive and possess source, special nuclear, or byproduct material at a geologic repository operations area, the DOE may (i) make changes in the geologic repository operations area as described in the...

  20. Results of instrument reliability study for high-level nuclear-waste repositories. [Geotechnical parameters

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rogue, F.; Binnall, E.P.

    1982-10-01

    Reliable instrumentation will be needed to monitor the performance of future high-level waste repository sites. A study has been made to assess instrument reliability at Department of Energy (DOE) waste repository related experiments. Though the study covers a wide variety of instrumentation, this paper concentrates on experiences with geotechnical instrumentation in hostile repository-type environments. Manufacturers have made some changes to improve the reliability of instruments for repositories. This paper reviews the failure modes, rates, and mechanisms, along with manufacturer modifications and recommendations for additional improvements to enhance instrument performance. 4 tables.

  1. Data repositories for medical education research: issues and recommendations.

    PubMed

    Schwartz, Alan; Pappas, Cleo; Sandlow, Leslie J

    2010-05-01

    The authors explore issues surrounding digital repositories with the twofold intention of clarifying their creation, structure, content, and use, and considering the implementation of a global digital repository for medical education research data sets-an online site where medical education researchers would be encouraged to deposit their data in order to facilitate the reuse and reanalysis of the data by other researchers. By motivating data sharing and reuse, investigators, medical schools, and other stakeholders might see substantial benefits to their own endeavors and to the progress of the field of medical education.The authors review digital repositories in medicine, social sciences, and education, describe the contents and scope of repositories, and present extant examples. The authors describe the potential benefits of a medical education data repository and report results of a survey of the Society for Directors of Research in Medicine Education, in which participants responded to questions about data sharing and a potential data repository. Respondents strongly endorsed data sharing, with the caveat that principal investigators should choose whether or not to share data they collect. A large majority believed that a repository would benefit their unit and the field of medical education. Few reported using existing repositories. Finally, the authors consider challenges to the establishment of such a repository, including taxonomic organization, intellectual property concerns, human subjects protection, technological infrastructure, and evaluation standards. The authors conclude with recommendations for how a medical education data repository could be successfully developed.

  2. Bio-repository of post-clinical test samples at the national cancer center hospital (NCCH) in Tokyo.

    PubMed

    Furuta, Koh; Yokozawa, Karin; Takada, Takako; Kato, Hoichi

    2009-08-01

    We established the Bio-repository at the National Cancer Center Hospital in October 2002. The main purpose of this article is to show the importance and usefulness of a bio-repository of post-clinical test samples not only for translational cancer research but also for routine clinical oncology by introducing the experience of setting up such a facility. Our basic concept of a post-clinical test sample is not as left-over waste, but rather as frozen evidence of a patient's pathological condition at a particular point. We can decode, if not all, most of the laboratory data from a post-clinical test sample. As a result, the bio-repository is able to provide not only the samples, but potentially all related laboratory data upon request. The areas of sample coverage are the following: sera after routine blood tests; sera after cross-match tests for transfusion; serum or plasma submitted at a patient's clinically important time period by the physician; and samples collected by the individual investigator. The formats of stored samples are plasma or serum, dried blood spot (DBS) and buffy coat. So far, 150 218 plasmas or sera, 35 253 DBS and 536 buffy coats have been registered for our bio-repository system. We arranged to provide samples to various concerned parties under strict legal and ethical agreements. Although the number of the utilized samples was initially limited, the inquiries for sample utilization are now increasing steadily from both research and clinical sources. Further efforts to increase the benefits of the repository are intended.

  3. Calculating the quality of public high-throughput sequencing data to obtain a suitable subset for reanalysis from the Sequence Read Archive

    PubMed Central

    Nakazato, Takeru; Bono, Hidemasa

    2017-01-01

    Abstract It is important for public data repositories to promote the reuse of archived data. In the growing field of omics science, however, the increasing number of submissions of high-throughput sequencing (HTSeq) data to public repositories prevents users from choosing a suitable data set from among the large number of search results. Repository users need to be able to set a threshold to reduce the number of results to obtain a suitable subset of high-quality data for reanalysis. We calculated the quality of sequencing data archived in a public data repository, the Sequence Read Archive (SRA), by using the quality control software FastQC. We obtained quality values for 1 171 313 experiments, which can be used to evaluate the suitability of data for reuse. We also visualized the data distribution in SRA by integrating the quality information and metadata of experiments and samples. We provide quality information of all of the archived sequencing data, which enable users to obtain sufficient quality sequencing data for reanalyses. The calculated quality data are available to the public in various formats. Our data also provide an example of enhancing the reuse of public data by adding metadata to published research data by a third party. PMID:28449062

  4. Core Certification of Data Repositories: Trustworthiness and Long-Term Stewardship

    NASA Astrophysics Data System (ADS)

    de Sherbinin, A. M.; Mokrane, M.; Hugo, W.; Sorvari, S.; Harrison, S.

    2017-12-01

    Scientific integrity and norms dictate that data created and used by scientists should be managed, curated, and archived in trustworthy data repositories thus ensuring that science is verifiable and reproducible while preserving the initial investment in collecting data. Research stakeholders including researchers, science funders, librarians, and publishers must also be able to establish the trustworthiness of data repositories they use to confirm that the data they submit and use remain useful and meaningful in the long term. Data repositories are increasingly recognized as a key element of the global research infrastructure and the importance of establishing their trustworthiness is recognised as a prerequisite for efficient scientific research and data sharing. The Core Trustworthy Data Repository Requirements are a set of universal requirements for certification of data repositories at the core level (see: https://goo.gl/PYsygW). They were developed by the ICSU World Data System (WDS: www.icsu-wds.org) and the Data Seal of Approval (DSA: www.datasealofapproval.org)—the two authoritative organizations responsible for the development and implementation of this standard to be further developed under the CoreTrustSeal branding . CoreTrustSeal certification of data repositories involves a minimally intensive process whereby repositories supply evidence that they are sustainable and trustworthy. Repositories conduct a self-assessment which is then reviewed by community peers. Based on this review CoreTrustSeal certification is granted by the CoreTrustSeal Standards and Certification Board. Certification helps data communities—producers, repositories, and consumers—to improve the quality and transparency of their processes, and to increase awareness of and compliance with established standards. This presentation will introduce the CoreTrustSeal certification requirements for repositories and offer an opportunity to discuss ways to improve the contribution of certified data repositories to sustain open data for open scientific research.

  5. Institutional Repositories in Indian Universities and Research Institutes: A Study

    ERIC Educational Resources Information Center

    Krishnamurthy, M.; Kemparaju, T. D.

    2011-01-01

    Purpose: The purpose of this paper is to report on a study of the institutional repositories (IRs) in use in Indian universities and research institutes. Design/methodology/approach: Repositories in various institutions in India were accessed and described in a standardised way. Findings: The 20 repositories studied covered collections of diverse…

  6. Repositories for Research: Southampton's Evolving Role in the Knowledge Cycle

    ERIC Educational Resources Information Center

    Simpson, Pauline; Hey, Jessie

    2006-01-01

    Purpose: To provide an overview of how open access (OA) repositories have grown to take a premier place in the e-research knowledge cycle and offer Southampton's route from project to sustainable institutional repository. Design/methodology/approach: The evolution of institutional repositories and OA is outlined raising questions of multiplicity…

  7. Reproducible Research in the Geosciences at Scale: Achievable Goal or Elusive Dream?

    NASA Astrophysics Data System (ADS)

    Wyborn, L. A.; Evans, B. J. K.

    2016-12-01

    Reproducibility is a fundamental tenant of the scientific method: it implies that any researcher, or a third party working independently, can duplicate any experiment or investigation and produce the same results. Historically computationally based research involved an individual using their own data and processing it in their own private area, often using software they wrote or inherited from close collaborators. Today, a researcher is likely to be part of a large team that will use a subset of data from an external repository and then process the data on a public or private cloud or on a large centralised supercomputer, using a mixture of their own code, third party software and libraries, or global community codes. In 'Big Geoscience' research it is common for data inputs to be extracts from externally managed dynamic data collections, where new data is being regularly appended, or existing data is revised when errors are detected and/or as processing methods are improved. New workflows increasingly use services to access data dynamically to create subsets on-the-fly from distributed sources, each of which can have a complex history. At major computational facilities, underlying systems, libraries, software and services are being constantly tuned and optimised, or as new or replacement infrastructure being installed. Likewise code used from a community repository is continually being refined, re-packaged and ported to the target platform. To achieve reproducibility, today's researcher increasingly needs to track their workflow, including querying information on the current or historical state of facilities used. Versioning methods are standard practice for software repositories or packages, but it is not common for either data repositories or data services to provide information about their state, or for systems to provide query-able access to changes in the underlying software. While a researcher can achieve transparency and describe steps in their workflow so that others can repeat them and replicate processes undertaken, they cannot achieve exact reproducibility or even transparency of results generated. In Big Geoscience, full reproducibiliy will be an elusive dream until data repositories and compute facilities can provide provenance information in a standards compliant, machine query-able way.

  8. New Features of the re3data Registry of Research Data Repositories

    NASA Astrophysics Data System (ADS)

    Elger, K.; Pampel, H.; Vierkant, P.; Witt, M.

    2016-12-01

    re3data is a registry of research data repositories that lists over 1,600 repositories from around the world, making it the largest and most comprehensive online catalog of data repositories on the web. The registry offers researchers, funding agencies, libraries and publishers a comprehensive overview of the heterogeneous landscape of data repositories. The repositories are described, following the "Metadata Schema for the Description of Research Data Repositories". re3data summarises the properties of a repository into a user-friendly icon system helping users to easily identify an adequate repository for the storage of their datasets. The re3data entries are curated by an international, multi-disciplinary editorial board. An application programming interface (API) enables other information systems to list and fetch metadata for integration and interoperability. Funders like the European Commission (2015) and publishers like Springer Nature (2016) recommend the use of re3data.org in their policies. The original re3data project partners are the GFZ German Research Centre for Geosciences, the Humboldt-Universität zu Berlin, the Purdue University Libraries and the Karlsruhe Institute of Technology (KIT). Since 2015 re3data is operated as a service of DataCite, a global non-profit organisation that provides persistent identifiers (DOIs) for research data. At the 2016 AGU Fall Meeting we will describe the current status of re3data. An overview of the major developments and new features will be given. Furthermore, we will present our plans to increase the quality of the re3data entries.

  9. Combining partially ranked data in plant breeding and biology: II. Analysis with Rasch model.

    USDA-ARS?s Scientific Manuscript database

    Many years of breeding experiments, germplasm screening, and molecular biologic experimentation have generated volumes of sequence, genotype, and phenotype information that have been stored in public data repositories. These resources afford genetic and genomic researchers the opportunity to handle ...

  10. Calculating the quality of public high-throughput sequencing data to obtain a suitable subset for reanalysis from the Sequence Read Archive.

    PubMed

    Ohta, Tazro; Nakazato, Takeru; Bono, Hidemasa

    2017-06-01

    It is important for public data repositories to promote the reuse of archived data. In the growing field of omics science, however, the increasing number of submissions of high-throughput sequencing (HTSeq) data to public repositories prevents users from choosing a suitable data set from among the large number of search results. Repository users need to be able to set a threshold to reduce the number of results to obtain a suitable subset of high-quality data for reanalysis. We calculated the quality of sequencing data archived in a public data repository, the Sequence Read Archive (SRA), by using the quality control software FastQC. We obtained quality values for 1 171 313 experiments, which can be used to evaluate the suitability of data for reuse. We also visualized the data distribution in SRA by integrating the quality information and metadata of experiments and samples. We provide quality information of all of the archived sequencing data, which enable users to obtain sufficient quality sequencing data for reanalyses. The calculated quality data are available to the public in various formats. Our data also provide an example of enhancing the reuse of public data by adding metadata to published research data by a third party. © The Authors 2017. Published by Oxford University Press.

  11. NASA Life Sciences Data Repositories: Tools for Retrospective Analysis and Future Planning

    NASA Technical Reports Server (NTRS)

    Thomas, D.; Wear, M.; VanBaalen, M.; Lee, L.; Fitts, M.

    2011-01-01

    As NASA transitions from the Space Shuttle era into the next phase of space exploration, the need to ensure the capture, analysis, and application of its research and medical data is of greater urgency than at any other previous time. In this era of limited resources and challenging schedules, the Human Research Program (HRP) based at NASA s Johnson Space Center (JSC) recognizes the need to extract the greatest possible amount of information from the data already captured, as well as focus current and future research funding on addressing the HRP goal to provide human health and performance countermeasures, knowledge, technologies, and tools to enable safe, reliable, and productive human space exploration. To this end, the Science Management Office and the Medical Informatics and Health Care Systems Branch within the HRP and the Space Medicine Division have been working to make both research data and clinical data more accessible to the user community. The Life Sciences Data Archive (LSDA), the research repository housing data and information regarding the physiologic effects of microgravity, and the Lifetime Surveillance of Astronaut Health (LSAH-R), the clinical repository housing astronaut data, have joined forces to achieve this goal. The task of both repositories is to acquire, preserve, and distribute data and information both within the NASA community and to the science community at large. This is accomplished via the LSDA s public website (http://lsda.jsc.nasa.gov), which allows access to experiment descriptions including hardware, datasets, key personnel, mission descriptions and a mechanism for researchers to request additional data, research and clinical, that is not accessible from the public website. This will result in making the work of NASA and its partners available to the wider sciences community, both domestic and international. The desired outcome is the use of these data for knowledge discovery, retrospective analysis, and planning of future research studies.

  12. Classification of Clinical Research Study Eligibility Criteria to Support Multi-Stage Cohort Identification Using Clinical Data Repositories.

    PubMed

    Cimino, James J; Lancaster, William J; Wyatt, Mathew C

    2017-01-01

    One of the challenges to using electronic health record (EHR) repositories for research is the difficulty mapping study subject eligibility criteria to the query capabilities of the repository. We sought to characterize criteria as "easy" (searchable in a typical repository), "hard" (requiring manual review of the record data), and "impossible" (not typically available in EHR repositories). We obtained 292 criteria from 20 studies available from Clinical Trials.gov and rated them according to our three types, plus a fourth "mixed" type. We had good agreement among three independent reviewers and chose 274 criteria that were characterized by single types for further analysis. The resulting analysis showed typical features of criteria that do and don't map to repositories. We propose that these features be used to guide researchers in specifying eligibility criteria to improve development of enrollment workflow, including the definition of EHR repository queries for self-service or analyst-mediated retrievals.

  13. Implementing and Sustaining Data Lifecycle best Practices: a Framework for Researchers and Repositories

    NASA Astrophysics Data System (ADS)

    Stall, S.

    2016-02-01

    Emerging data management mandates in conjunction with cross-domain international interoperability are posing new challenges for researchers and repositories. Domain repositories are serving in this critical, growing role monitoring and leading data management standards and capability within their own repository and working on mappings between repositories internationally. Leading research institutions and companies will also be important as they develop and expand data curation efforts. This landscape poses a number of challenges for developing and ensuring the use of best practices in curating research data, enabling discovery, elevating quality across diverse repositories, and helping researchers collect and organize it through the full data life cycle. This multidimensional challenge will continue to grow in complexity. The American Geophysical Union (AGU) is developing two programs to help researchers and data repositories develop and elevate best practices and address these challenges. The goal is to provide tools for the researchers and repositories, whether domain, institutional, or other, that improve performance throughout the data lifecycle across the Earth and space science community. For scientists and researchers, AGU is developing courses around handling data that can lead toward a certification in geoscience data management. Course materials will cover metadata management and collection, data analysis, integration of data, and data presentation. The course topics are being finalized by the advisory board with the first one planned to be available later this year. AGU is also developing a program aimed at helping data repositories, large and small, domain-specific to general, assess and improve data management practices. AGU has partnered with the CMMI® Institute to adapt their Data Management Maturity (DMM)SM framework within the Earth and space sciences. A data management assessment using the DMMSM involves identifying accomplishments and weaknesses compared to leading practices for data management. Recommendations can help improve quality and consistency across the community that will facilitate reuse in the data lifecycle. Through governance, quality, and architecture process areas the assessment can measure the ability for data to be discoverable and interoperable.

  14. Introducing the Brassica Information Portal: Towards integrating genotypic and phenotypic Brassica crop data

    PubMed Central

    Eckes, Annemarie H.; Gubała, Tomasz; Nowakowski, Piotr; Szymczyszyn, Tomasz; Wells, Rachel; Irwin, Judith A.; Horro, Carlos; Hancock, John M.; King, Graham; Dyer, Sarah C.; Jurkowski, Wiktor

    2017-01-01

    The Brassica Information Portal (BIP) is a centralised repository for brassica phenotypic data. The site hosts trait data associated with brassica research and breeding experiments conducted on brassica crops, that are used as oilseeds, vegetables, livestock forage and fodder and for biofuels. A key feature is the explicit management of meta-data describing the provenance and relationships between experimental plant materials, as well as trial design and trait descriptors. BIP is an open access and open source project, built on the schema of CropStoreDB, and as such can provide trait data management strategies for any crop data. A new user interface and programmatic submission/retrieval system helps to simplify data access for researchers, breeders and other end-users. BIP opens up the opportunity to apply integrative, cross-project analyses to data generated by the Brassica Research Community. Here, we present a short description of the current status of the repository. PMID:28529710

  15. ArrayWiki: an enabling technology for sharing public microarray data repositories and meta-analyses

    PubMed Central

    Stokes, Todd H; Torrance, JT; Li, Henry; Wang, May D

    2008-01-01

    Background A survey of microarray databases reveals that most of the repository contents and data models are heterogeneous (i.e., data obtained from different chip manufacturers), and that the repositories provide only basic biological keywords linking to PubMed. As a result, it is difficult to find datasets using research context or analysis parameters information beyond a few keywords. For example, to reduce the "curse-of-dimension" problem in microarray analysis, the number of samples is often increased by merging array data from different datasets. Knowing chip data parameters such as pre-processing steps (e.g., normalization, artefact removal, etc), and knowing any previous biological validation of the dataset is essential due to the heterogeneity of the data. However, most of the microarray repositories do not have meta-data information in the first place, and do not have a a mechanism to add or insert this information. Thus, there is a critical need to create "intelligent" microarray repositories that (1) enable update of meta-data with the raw array data, and (2) provide standardized archiving protocols to minimize bias from the raw data sources. Results To address the problems discussed, we have developed a community maintained system called ArrayWiki that unites disparate meta-data of microarray meta-experiments from multiple primary sources with four key features. First, ArrayWiki provides a user-friendly knowledge management interface in addition to a programmable interface using standards developed by Wikipedia. Second, ArrayWiki includes automated quality control processes (caCORRECT) and novel visualization methods (BioPNG, Gel Plots), which provide extra information about data quality unavailable in other microarray repositories. Third, it provides a user-curation capability through the familiar Wiki interface. Fourth, ArrayWiki provides users with simple text-based searches across all experiment meta-data, and exposes data to search engine crawlers (Semantic Agents) such as Google to further enhance data discovery. Conclusions Microarray data and meta information in ArrayWiki are distributed and visualized using a novel and compact data storage format, BioPNG. Also, they are open to the research community for curation, modification, and contribution. By making a small investment of time to learn the syntax and structure common to all sites running MediaWiki software, domain scientists and practioners can all contribute to make better use of microarray technologies in research and medical practices. ArrayWiki is available at . PMID:18541053

  16. 2012 best practices for repositories collection, storage, retrieval, and distribution of biological materials for research international society for biological and environmental repositories.

    PubMed

    2012-04-01

    Third Edition [Formula: see text] [Box: see text] Printed with permission from the International Society for Biological and Environmental Repositories (ISBER) © 2011 ISBER All Rights Reserved Editor-in-Chief Lori D. Campbell, PhD Associate Editors Fay Betsou, PhD Debra Leiolani Garcia, MPA Judith G. Giri, PhD Karen E. Pitt, PhD Rebecca S. Pugh, MS Katherine C. Sexton, MBA Amy P.N. Skubitz, PhD Stella B. Somiari, PhD Individual Contributors to the Third Edition Jonas Astrin, Susan Baker, Thomas J. Barr, Erica Benson, Mark Cada, Lori Campbell, Antonio Hugo Jose Froes Marques Campos, David Carpentieri, Omoshile Clement, Domenico Coppola, Yvonne De Souza, Paul Fearn, Kelly Feil, Debra Garcia, Judith Giri, William E. Grizzle, Kathleen Groover, Keith Harding, Edward Kaercher, Joseph Kessler, Sarah Loud, Hannah Maynor, Kevin McCluskey, Kevin Meagher, Cheryl Michels, Lisa Miranda, Judy Muller-Cohn, Rolf Muller, James O'Sullivan, Karen Pitt, Rebecca Pugh, Rivka Ravid, Katherine Sexton, Ricardo Luis A. Silva, Frank Simione, Amy Skubitz, Stella Somiari, Frans van der Horst, Gavin Welch, Andy Zaayenga 2012 Best Practices for Repositories: Collection, Storage, Retrieval and Distribution of Biological Materials for Research INTERNATIONAL SOCIETY FOR BIOLOGICAL AND ENVIRONMENTAL REPOSITORIES (ISBER) INTRODUCTION T he availability of high quality biological and environmental specimens for research purposes requires the development of standardized methods for collection, long-term storage, retrieval and distribution of specimens that will enable their future use. Sharing successful strategies for accomplishing this goal is one of the driving forces for the International Society for Biological and Environmental Repositories (ISBER). For more information about ISBER see www.isber.org . ISBER's Best Practices for Repositories (Best Practices) reflect the collective experience of its members and has received broad input from other repository professionals. Throughout this document effective practices are presented for the management of specimen collections and repositories. The term "Best Practice" is used in cases where a level of operation is indicated that is above the basic recommended practice or more specifically designates the most effective practice. It is understood that repositories in certain locations or with particular financial constraints may not be able to adhere to each of the items designated as "Best Practices". Repositories fitting into either of these categories will need to decide how they might best adhere to these recommendations within their particular circumstances. While adherence to ISBER Best Practices is strictly on a voluntary basis, it is important to note that some aspects of specimen management are governed by national/federal, regional and local regulations. The reader should refer directly to regulations for their national/federal, regional and local requirements, as appropriate. ISBER has strived to include terminology appropriate to the various specimen types covered under these practices, but here too, the reader should take steps to ensure the appropriateness of the recommendations to their particular repository type prior to the implementation of any new approaches. Important terms within the document are italicized when first used in a section and defined in the glossary. The ISBER Best Practices are periodically reviewed and revised to reflect advances in research and technology. The third edition of the Best Practices builds on the foundation established in the first and second editions which were published in 2005 and 2008, respectively.

  17. Making the Most of What We Have Got: Enhancing the RADAR Repository to Support Research Planning

    ERIC Educational Resources Information Center

    Pike, Dawn; Siminson, Nicola Jane

    2017-01-01

    This article discusses how RADAR, the institutional repository (IR) at the Glasgow School of Art (GSA), has been modified to house an Annual Research Planning (ARP) template. A case study on the implementation of this research planning tool will outline the role that a repository and its staff can play in supporting individuals, enhancing…

  18. NCI Mouse Repository | Frederick National Laboratory for Cancer Research

    Cancer.gov

    The NCI Mouse Repository is an NCI-funded resource for mouse cancer models and associated strains. The repository makes strains available to all members of the scientific community (academic, non-profit, and commercial). NCI Mouse Repository strains

  19. Benefits of International Collaboration on the International Space Station

    NASA Technical Reports Server (NTRS)

    Hasbrook, Pete; Robinson, Julie A.; Cohen, Luchino; Marcil, Isabelle; De Parolis, Lina; Hatton, Jason; Shirakawa, Masaki; Karabadzhak, Georgy; Sorokin, Igor V.; Valentini, Giovanni

    2017-01-01

    The International Space Station is a valuable platform for research in space, but the benefits are limited if research is only conducted by individual countries. Through the efforts of the ISS Program Science Forum, international science working groups, and interagency cooperation, international collaboration on the ISS has expanded as ISS utilization has matured. Members of science teams benefit from working with counterparts in other countries. Scientists and institutions bring years of experience and specialized expertise to collaborative investigations, leading to new perspectives and approaches to scientific challenges. Combining new ideas and historical results brings synergy and improved peer-reviewed scientific methods and results. World-class research facilities can be expensive and logistically complicated, jeopardizing their full utilization. Experiments that would be prohibitively expensive for a single country can be achieved through contributions of resources from two or more countries, such as crew time, up- and down mass, and experiment hardware. Cooperation also avoids duplication of experiments and hardware among agencies. Biomedical experiments can be completed earlier if astronauts or cosmonauts from multiple agencies participate. Countries responding to natural disasters benefit from ISS imagery assets, even if the country has no space agency of its own. Students around the world participate in ISS educational opportunities, and work with students in other countries, through open curriculum packages and through international competitions. Even experiments conducted by a single country can benefit scientists around the world, through specimen sharing programs and publicly accessible "open data" repositories. For ISS data, these repositories include GeneLab, the Physical Science Informatics System, and different Earth data systems. Scientists can conduct new research using ISS data without having to launch and execute their own experiments. Multilateral collections of research results publications, maintained by the ISS international partnership and accessible via nasa.gov, make ISS results available worldwide, and encourage new users, ideas and research.

  20. Creation of Data Repositories to Advance Nursing Science.

    PubMed

    Perazzo, Joseph; Rodriguez, Margaret; Currie, Jackson; Salata, Robert; Webel, Allison R

    2017-12-01

    Data repositories are a strategy in line with precision medicine and big data initiatives, and are an efficient way to maximize data utility and form collaborative research relationships. Nurse researchers are uniquely positioned to make a valuable contribution using this strategy. The purpose of this article is to present a review of the benefits and challenges associated with developing data repositories, and to describe the process we used to develop and maintain a data repository in HIV research. Systematic planning, data collection, synthesis, and data sharing have enabled us to conduct robust cross-sectional and longitudinal analyses with more than 200 people living with HIV. Our repository building has also led to collaboration and training, both in and out of our organization. We present a pragmatic and affordable way that nurse scientists can build and maintain a data repository, helping us continue to make to our understanding of health phenomena.

  1. re3data.org - a global registry of research data repositories

    NASA Astrophysics Data System (ADS)

    Pampel, Heinz; Vierkant, Paul; Elger, Kirsten; Bertelmann, Roland; Witt, Michael; Schirmbacher, Peter; Rücknagel, Jessika; Kindling, Maxi; Scholze, Frank; Ulrich, Robert

    2016-04-01

    re3data.org - the registry of research data repositories lists over 1,400 research data repositories from all over the world making it the largest and most comprehensive online catalog of research data repositories on the web. The registry is a valuable tool for researchers, funding organizations, publishers and libraries. re3data.org provides detailed information about research data repositories, and its distinctive icons help researchers to easily identify relevant repositories for accessing and depositing data sets [1]. Funding agencies, like the European Commission [2] and research institutions like the University of Bielefeld [3] already recommend the use of re3data.org in their guidelines and policies. Several publishers and journals like Copernicus Publications, PeerJ, and Nature's Scientific Data recommend re3data.org in their editorial policies as a tool for the easy identification of appropriate data repositories to store research data. Project partners in re3data.org are the Library and Information Services department (LIS) of the GFZ German Research Centre for Geosciences, the Computer and Media Service at the Humboldt-Universität zu Berlin, the Purdue University Libraries and the KIT Library at the Karlsruhe Institute of Technology (KIT). After its fusion with the U.S. American DataBib in 2014, re3data.org continues as a service of DataCite from 2016 on. DataCite is the international organization for the registration of Digital Object Identifiers (DOI) for research data and aims to improve their citation. The poster describes the current status and the future plans of re3data.org. [1] Pampel H, et al. (2013) Making Research Data Repositories Visible: The re3data.org Registry. PLoS ONE 8(11): e78080. doi:10.1371/journal.pone.0078080. [2] European Commission (2015): Guidelines on Open Access to Scientific Publications and Research Data in Horizon 2020. Available: http://ec.europa.eu/research/participants/data/ref/h2020/grants_manual/hi/oa_pilot/h2020-hi-oa-pilot-guide_en.pdf Accessed 11 January 2016. [3] Bielefeld University (2013): Resolution on Research Data Management. Available: http://data.uni-bielefeld.de/en/resolution Accessed 11 January 2016.

  2. Legal Agreements and the Governance of Research Commons: Lessons from Materials Sharing in Mouse Genomics

    PubMed Central

    Mishra, Amrita

    2014-01-01

    Abstract Omics research infrastructure such as databases and bio-repositories requires effective governance to support pre-competitive research. Governance includes the use of legal agreements, such as Material Transfer Agreements (MTAs). We analyze the use of such agreements in the mouse research commons, including by two large-scale resource development projects: the International Knockout Mouse Consortium (IKMC) and International Mouse Phenotyping Consortium (IMPC). We combine an analysis of legal agreements and semi-structured interviews with 87 members of the mouse model research community to examine legal agreements in four contexts: (1) between researchers; (2) deposit into repositories; (3) distribution by repositories; and (4) exchanges between repositories, especially those that are consortium members of the IKMC and IMPC. We conclude that legal agreements for the deposit and distribution of research reagents should be kept as simple and standard as possible, especially when minimal enforcement capacity and resources exist. Simple and standardized legal agreements reduce transactional bottlenecks and facilitate the creation of a vibrant and sustainable research commons, supported by repositories and databases. PMID:24552652

  3. Functional requirements for a central research imaging data repository.

    PubMed

    Franke, Thomas; Gruetz, Romanus; Dickmann, Frank

    2013-01-01

    The current situation at many university medical centers regarding the management of biomedical research imaging data leaves much to be desired. In contrast to the recommendations of the German Research Foundation (DFG) and the German Council of Sciences and Humanities regarding the professional management of research data, there are commonly many individual data pools for research data in each institute and the management remains the responsibility of the researcher. A possible solution for this situation would be to install local central repositories for biomedical research imaging data. In this paper, we developed a scenario based on abstracted use-cases for institutional research undertakings as well as collaborative biomedical research projects and analyzed the functional requirements that a local repository would have to fulfill. We determined eight generic categories of functional requirements, which can be viewed as a basic guideline for the minimum functionality of a central repository for biomedical research imaging data.

  4. Desiderata for Healthcare Integrated Data Repositories Based on Architectural Comparison of Three Public Repositories

    PubMed Central

    Huser, Vojtech; Cimino, James J.

    2013-01-01

    Integrated data repositories (IDRs) are indispensable tools for numerous biomedical research studies. We compare three large IDRs (Informatics for Integrating Biology and the Bedside (i2b2), HMO Research Network’s Virtual Data Warehouse (VDW) and Observational Medical Outcomes Partnership (OMOP) repository) in order to identify common architectural features that enable efficient storage and organization of large amounts of clinical data. We define three high-level classes of underlying data storage models and we analyze each repository using this classification. We look at how a set of sample facts is represented in each repository and conclude with a list of desiderata for IDRs that deal with the information storage model, terminology model, data integration and value-sets management. PMID:24551366

  5. Desiderata for healthcare integrated data repositories based on architectural comparison of three public repositories.

    PubMed

    Huser, Vojtech; Cimino, James J

    2013-01-01

    Integrated data repositories (IDRs) are indispensable tools for numerous biomedical research studies. We compare three large IDRs (Informatics for Integrating Biology and the Bedside (i2b2), HMO Research Network's Virtual Data Warehouse (VDW) and Observational Medical Outcomes Partnership (OMOP) repository) in order to identify common architectural features that enable efficient storage and organization of large amounts of clinical data. We define three high-level classes of underlying data storage models and we analyze each repository using this classification. We look at how a set of sample facts is represented in each repository and conclude with a list of desiderata for IDRs that deal with the information storage model, terminology model, data integration and value-sets management.

  6. Introducing Ocean Science Research to Two-Year College (2YC) Students Through Inquiry-Based Experiences

    NASA Astrophysics Data System (ADS)

    Gamage, K. R.

    2016-02-01

    An effective approach to introduce 2YC students to ocean science research is through propagating inquiry-based experiences into existing geosciences courses using a series of research activities. The proposed activity is based on scientific ocean drilling, where students begin their research experience (pre-field activity) by reading articles from scientific journals and analyzing and interpreting core and log data on a specific research topic. At the end of the pre-field activity, students will visit the Gulf Coast Repository to examine actual cores, smear slides, thin sections etc. After the visit, students will integrate findings from their pre-field and field activities to produce a term paper. These simple activities allow students to experience in the iterative process of scientific research, illuminates how scientists approach ocean science, and can be the hook to get students interested in pursuing ocean science as a career.

  7. US/German Collaboration in Salt Repository Research, Design and Operation - 13243

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Steininger, Walter; Hansen, Frank; Biurrun, Enrique

    2013-07-01

    Recent developments in the US and Germany [1-3] have precipitated renewed efforts in salt repository investigations and related studies. Both the German rock salt repository activities and the US waste management programs currently face challenges that may adversely affect their respective current and future state-of-the-art core capabilities in rock salt repository science and technology. The research agenda being pursued by our respective countries leverages collective efforts for the benefit of both programs. The topics addressed by the US/German salt repository collaborations align well with the findings and recommendations summarized in the January 2012 US Blue Ribbon Commission on America's Nuclearmore » Future (BRC) report [4] and are consistent with the aspirations of the key topics of the Strategic Research Agenda of the Implementing Geological Disposal of Radioactive Waste Technology Platform (IGD-TP) [5]. Against this background, a revival of joint efforts in salt repository investigations after some years of hibernation has been undertaken to leverage collective efforts in salt repository research, design, operations, and related issues for the benefit of respective programs and to form a basis for providing an attractive, cost-effective insurance against the premature loss of virtually irreplaceable scientific expertise and institutional memory. (authors)« less

  8. Community based research for an urban recreation application of benefits-based management

    Treesearch

    William T. Borrie; Joseph W. Roggenbuck

    1995-01-01

    Benefits-based management is an approach to park and recreation management that focuses on the positive outcomes of engaging in recreational experiences. Because one class of possible benefits accrue to the community, a philosophical framework is discussed suggesting that communities are themselves the primary sources, generators, and repositories of knowledge....

  9. Large-Scale In-situ Experiments to Determine Geochemical Alterations and Microbial Activities at the Geological Repository

    NASA Astrophysics Data System (ADS)

    Choung, S.; Francis, A. J.; Um, W.; Choi, S.; Kim, S.; Park, J.; Kim, S.

    2013-12-01

    The countries that have generated nuclear power have facing problems on the disposal of accumulated radioactive wastes. Geological disposal method has been chosen in many countries including Korea. A safety issue after the closure of geological repository has been raised, because microbial activities lead overpressure in the underground facilities through gas production. In particular, biodegradable organic materials derived from low- and intermediate-level radioactive wastes play important role on microbial activities in the geological repository. This study performed large scale in-situ experiments using organic wastes and groundwater, and investigated geochemical alteration and microbial activities at early stage (~63 days) as representative of the period, after closure of the geological repository. The geochemical alteration controlled significantly the microorganism types and populations. Database of the biogeochemical alteration facilitates prediction of radionuclides' mobility and establishment of remedial strategy against unpredictable accidents and hazards at early stage right after closure of the geological repository.

  10. Science is the first step to siting nuclear waste repositories

    USGS Publications Warehouse

    Neuzil, Christopher E.

    2014-01-01

    As Shaw [2014] notes, U.S. research on shale as a repository host was halted before expending anything close to the effort devoted to studying crystalline rock, salt, and - most notably - tuff at Yucca Mountain. The new political reality regarding Yucca Mountain may allow reconsideration of the decision to abandon research on shale as a repository host.

  11. Virtual patient repositories--a comparative analysis.

    PubMed

    Küfner, Julia; Kononowicz, Andrzej A; Hege, Inga

    2014-01-01

    Virtual Patients (VPs) are an important component of medical education. One way to reduce the costs for creating VPs is sharing through repositories. We conducted a literature review to identify existing repositories and analyzed the 17 included repositories in regards to the search functions and metadata they provide. Most repositories provided some metadata such as title or description, whereas other data, such as educational objectives, were less frequent. Future research could, in cooperation with the repository provider, investigate user expectations and usage patterns.

  12. International Collaboration Activities on Engineered Barrier Systems

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jove-Colon, Carlos F.

    The Used Fuel Disposition Campaign (UFDC) within the DOE Fuel Cycle Technologies (FCT) program has been engaging in international collaborations between repository R&D programs for high-level waste (HLW) disposal to leverage on gathered knowledge and laboratory/field data of near- and far-field processes from experiments at underground research laboratories (URL). Heater test experiments at URLs provide a unique opportunity to mimetically study the thermal effects of heat-generating nuclear waste in subsurface repository environments. Various configurations of these experiments have been carried out at various URLs according to the disposal design concepts of the hosting country repository program. The FEBEX (Full-scale Engineeredmore » Barrier Experiment in Crystalline Host Rock) project is a large-scale heater test experiment originated by the Spanish radioactive waste management agency (Empresa Nacional de Residuos Radiactivos S.A. – ENRESA) at the Grimsel Test Site (GTS) URL in Switzerland. The project was subsequently managed by CIEMAT. FEBEX-DP is a concerted effort of various international partners working on the evaluation of sensor data and characterization of samples obtained during the course of this field test and subsequent dismantling. The main purpose of these field-scale experiments is to evaluate feasibility for creation of an engineered barrier system (EBS) with a horizontal configuration according to the Spanish concept of deep geological disposal of high-level radioactive waste in crystalline rock. Another key aspect of this project is to improve the knowledge of coupled processes such as thermal-hydro-mechanical (THM) and thermal-hydro-chemical (THC) operating in the near-field environment. The focus of these is on model development and validation of predictions through model implementation in computational tools to simulate coupled THM and THC processes.« less

  13. SHIWA Services for Workflow Creation and Sharing in Hydrometeorolog

    NASA Astrophysics Data System (ADS)

    Terstyanszky, Gabor; Kiss, Tamas; Kacsuk, Peter; Sipos, Gergely

    2014-05-01

    Researchers want to run scientific experiments on Distributed Computing Infrastructures (DCI) to access large pools of resources and services. To run these experiments requires specific expertise that they may not have. Workflows can hide resources and services as a virtualisation layer providing a user interface that researchers can use. There are many scientific workflow systems but they are not interoperable. To learn a workflow system and create workflows may require significant efforts. Considering these efforts it is not reasonable to expect that researchers will learn new workflow systems if they want to run workflows developed in other workflow systems. To overcome it requires creating workflow interoperability solutions to allow workflow sharing. The FP7 'Sharing Interoperable Workflow for Large-Scale Scientific Simulation on Available DCIs' (SHIWA) project developed the Coarse-Grained Interoperability concept (CGI). It enables recycling and sharing workflows of different workflow systems and executing them on different DCIs. SHIWA developed the SHIWA Simulation Platform (SSP) to implement the CGI concept integrating three major components: the SHIWA Science Gateway, the workflow engines supported by the CGI concept and DCI resources where workflows are executed. The science gateway contains a portal, a submission service, a workflow repository and a proxy server to support the whole workflow life-cycle. The SHIWA Portal allows workflow creation, configuration, execution and monitoring through a Graphical User Interface using the WS-PGRADE workflow system as the host workflow system. The SHIWA Repository stores the formal description of workflows and workflow engines plus executables and data needed to execute them. It offers a wide-range of browse and search operations. To support non-native workflow execution the SHIWA Submission Service imports the workflow and workflow engine from the SHIWA Repository. This service either invokes locally or remotely pre-deployed workflow engines or submits workflow engines with the workflow to local or remote resources to execute workflows. The SHIWA Proxy Server manages certificates needed to execute the workflows on different DCIs. Currently SSP supports sharing of ASKALON, Galaxy, GWES, Kepler, LONI Pipeline, MOTEUR, Pegasus, P-GRADE, ProActive, Triana, Taverna and WS-PGRADE workflows. Further workflow systems can be added to the simulation platform as required by research communities. The FP7 'Building a European Research Community through Interoperable Workflows and Data' (ER-flow) project disseminates the achievements of the SHIWA project to build workflow user communities across Europe. ER-flow provides application supports to research communities within (Astrophysics, Computational Chemistry, Heliophysics and Life Sciences) and beyond (Hydrometeorology and Seismology) to develop, share and run workflows through the simulation platform. The simulation platform supports four usage scenarios: creating and publishing workflows in the repository, searching and selecting workflows in the repository, executing non-native workflows and creating and running meta-workflows. The presentation will outline the CGI concept, the SHIWA Simulation Platform, the ER-flow usage scenarios and how the Hydrometeorology research community runs simulations on SSP.

  14. Trustworthy Digital Repositories: Building Trust the Old Fashion Way, EARNING IT.

    NASA Astrophysics Data System (ADS)

    Kinkade, D.; Chandler, C. L.; Shepherd, A.; Rauch, S.; Groman, R. C.; Wiebe, P. H.; Glover, D. M.; Allison, M. D.; Copley, N. J.; Ake, H.; York, A.

    2016-12-01

    There are several drivers increasing the importance of high quality data management and curation in today's research process (e.g., OSTP PARR memo, journal publishers, funders, academic and private institutions), and proper management is necessary throughout the data lifecycle to enable reuse and reproducibility of results. Many digital data repositories are capable of satisfying the basic management needs of an investigator looking to share their data (i.e., publish data in the public domain), but repository services vary greatly and not all provide mature services that facilitate discovery, access, and reuse of research data. Domain-specific repositories play a vital role in the data curation process by working closely with investigators to create robust metadata, perform first order QC, and assemble and publish research data. In addition, they may employ technologies and services that enable increased discovery, access, and long-term archive. However, smaller domain facilities operate in varying states of capacity and curation ability. Within this repository environment, individual investigators (driven by publishers, funders, or institutions) need to find trustworthy repositories for their data; and funders need to direct investigators to quality repositories to ensure return on their investment. So, how can one determine the best home for valuable research data? Metrics can be applied to varying aspects of data curation, and many credentialing organizations offer services that assess and certify the trustworthiness of a given data management facility. Unfortunately, many of these certifications can be inaccessible to a small repository in cost, time, or scope. Are there alternatives? This presentation will discuss methods and approaches used by the Biological and Chemical Oceanography Data Management Office (BCO-DMO; a domain-specific, intermediate digital data repository) to demonstrate trustworthiness in the face of a daunting accreditation landscape.

  15. A data-rich recruitment core to support translational clinical research.

    PubMed

    Kost, Rhonda G; Corregano, Lauren M; Rainer, Tyler-Lauren; Melendez, Caroline; Coller, Barry S

    2015-04-01

    Underenrollment of clinical studies wastes resources and delays assessment of research discoveries. We describe the organization and impact of a centralized recruitment core delivering comprehensive recruitment support to investigators. The Rockefeller University Center for Clinical and Translational Science supports a centralized recruitment core, call center, Research Volunteer Repository, data infrastructure, and staff who provide expert recruitment services to investigators. During protocol development, consultations aim to optimize enrollment feasibility, develop recruitment strategy, budget, and advertising. Services during study conduct include advertising placement, repository queries, call management, prescreening, referral, and visit scheduling. Utilization and recruitment outcomes are tracked using dedicated software. For protocols receiving recruitment services during 2009-2013: median time from initiation of recruitment to the first enrolled participant was 10 days; of 4,047 first-time callers to the call center, 92% (n = 3,722) enrolled in the Research Volunteer Repository, with 99% retention; 23% of Repository enrollees subsequently enrolled in ≥1 research studies, with 89% retention. Of volunteers referred by repository queries, 49% (280/537) enrolled into the study, with 92% retained. Provision of robust recruitment infrastructure including expertise, a volunteer repository, data capture and real-time analysis accelerates protocol accrual. Application of recruitment science improves the quality of clinical investigation. © 2014 Wiley Periodicals, Inc.

  16. A Data‐Rich Recruitment Core to Support Translational Clinical Research

    PubMed Central

    Corregano, Lauren M.; Rainer, Tyler‐Lauren; Melendez, Caroline; Coller, Barry S.

    2014-01-01

    Abstract Background Underenrollment of clinical studies wastes resources and delays assessment of research discoveries. We describe the organization and impact of a centralized recruitment core delivering comprehensive recruitment support to investigators. Methods The Rockefeller University Center for Clinical and Translational Science supports a centralized recruitment core, call center, Research Volunteer Repository, data infrastructure, and staff who provide expert recruitment services to investigators. During protocol development, consultations aim to optimize enrollment feasibility, develop recruitment strategy, budget, and advertising. Services during study conduct include advertising placement, repository queries, call management, prescreening, referral, and visit scheduling. Utilization and recruitment outcomes are tracked using dedicated software. Results For protocols receiving recruitment services during 2009–2013: median time from initiation of recruitment to the first enrolled participant was 10 days; of 4,047 first‐time callers to the call center, 92% (n = 3,722) enrolled in the Research Volunteer Repository, with 99% retention; 23% of Repository enrollees subsequently enrolled in ≥1 research studies, with 89% retention. Of volunteers referred by repository queries, 49% (280/537) enrolled into the study, with 92% retained. Conclusions Provision of robust recruitment infrastructure including expertise, a volunteer repository, data capture and real‐time analysis accelerates protocol accrual. Application of recruitment science improves the quality of clinical investigation. PMID:25381717

  17. DataONE: Gateway to Earth and Environmental Data Repositories

    NASA Astrophysics Data System (ADS)

    Koskela, R.; Michener, W. K.; Vieglais, D.; Budden, A. E.

    2017-12-01

    DataONE (Data Observation Network for Earth) is a National Science Foundation DataNet project that enables universal access to data and also facilitates researchers in fulfilling their need for data management and in providing secure and permanent access to their data. DataONE offers the scientific community a suite of tools and training materials that cover all aspects of the data life cycle from data collection, to management, analysis and publication. Data repositories affiliated with DataONE are referred to as Member Nodes and represent large regional, national and international research networks, agencies, and other institutions. As part of the DataONE Federation, the repositories gain access to a range of value-added services to support their users. These services include usage tracking and reporting, content replication, and the ability to register the services created by the repository. In addition, DataONE and the California Digital Library manage ONEShare, a repository that accepts content submitted through Dash, a platform allowing researchers to easily describe, deposit and share their research data.

  18. The Use of Digital Repositories for Enhancing Teacher Pedagogical Performance

    ERIC Educational Resources Information Center

    Cohen, Anat; Kalimi, Sharon; Nachmias, Rafi

    2013-01-01

    This research examines the usage of local learning material repositories at school, as well as related teachers' attitudes and training. The study investigates the use of these repositories for enhancing teacher performance and assesses whether the assimilation of the local repositories increases their usage of and contribution to by teachers. One…

  19. The National Institute of Diabetes and Digestive and Kidney Diseases Central Repositories: A Valuable Resource for Nephrology Research

    PubMed Central

    Akolkar, Beena; Spain, Lisa M.; Guill, Michael H.; Del Vecchio, Corey T.; Carroll, Leslie E.

    2015-01-01

    The National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK) Central Repositories, part of the National Institutes of Health (NIH), are an important resource available to researchers and the general public. The Central Repositories house samples, genetic data, phenotypic data, and study documentation from >100 NIDDK-funded clinical studies, in areas such as diabetes, digestive disease, and liver disease research. The Central Repositories also have an exceptionally rich collection of studies related to kidney disease, including the Modification of Diet in Renal Disease landmark study and recent data from the Chronic Renal Insufficiency Cohort and CKD in Children Cohort studies. The data are carefully curated and linked to the samples from the study. The NIDDK is working to make the materials and data accessible to researchers. The Data Repositories continue to improve flexible online searching tools that help researchers identify the samples or data of interest, and NIDDK has created several different paths to access the data and samples, including some funding initiatives. Over the past several years, the Central Repositories have seen steadily increasing interest and use of the stored materials. NIDDK plans to make more collections available and do more outreach and education about use of the datasets to the nephrology research community in the future to enhance the value of this resource. PMID:25376765

  20. The National Institute of Diabetes and Digestive and Kidney Diseases Central Repositories: a valuable resource for nephrology research.

    PubMed

    Rasooly, Rebekah S; Akolkar, Beena; Spain, Lisa M; Guill, Michael H; Del Vecchio, Corey T; Carroll, Leslie E

    2015-04-07

    The National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK) Central Repositories, part of the National Institutes of Health (NIH), are an important resource available to researchers and the general public. The Central Repositories house samples, genetic data, phenotypic data, and study documentation from >100 NIDDK-funded clinical studies, in areas such as diabetes, digestive disease, and liver disease research. The Central Repositories also have an exceptionally rich collection of studies related to kidney disease, including the Modification of Diet in Renal Disease landmark study and recent data from the Chronic Renal Insufficiency Cohort and CKD in Children Cohort studies. The data are carefully curated and linked to the samples from the study. The NIDDK is working to make the materials and data accessible to researchers. The Data Repositories continue to improve flexible online searching tools that help researchers identify the samples or data of interest, and NIDDK has created several different paths to access the data and samples, including some funding initiatives. Over the past several years, the Central Repositories have seen steadily increasing interest and use of the stored materials. NIDDK plans to make more collections available and do more outreach and education about use of the datasets to the nephrology research community in the future to enhance the value of this resource. Copyright © 2015 by the American Society of Nephrology.

  1. The Challenges of Releasing Human Data for Analysis

    NASA Technical Reports Server (NTRS)

    Fitts, Mary; Van Baalen, Mary; Johnson-Throop, Kathy; Lee, Lesley; Havelka, Jacque; Wear, Mary; Thomas, Diedre M.

    2011-01-01

    The NASA Johnson Space Center s (NASA JSC) Committee for the Protection of Human Subjects (CPHS) recently approved the formation of two human data repositories: the Lifetime Surveillance of Astronaut Health Repository (LSAH-R) for clinical data and the Life Sciences Data Archive Repository (LSDA-R) for research data. The establishment of these repositories forms the foundation for the release of data and information beyond the scope for which the data was originally collected. The release of clinical and research data and information is primarily managed by two NASA groups: the Evidence Base Working Group (EBWG), consisting of members of both repositories, and the LSAH Policy Board. The goal of unifying these repositories and their processes is to provide a mutually supportive approach to handling medical and research data, to enhance the use of medical and research data to reduce risk, and to promote the understanding of space physiology, countermeasures and other mitigation strategies. Over the past year, both repositories have received over 100 data and information requests from a wide variety of requesters. The disposition of these requests has highlighted the challenges faced when attempting to make data collected on a unique set of subjects available beyond the original intent for which the data were collected. As the EBWG works through each request, many considerations must be factored into account when deciding what data can be shared and how - from the Privacy Act of 1974 and the Health Insurance Portability and Accountability Act (HIPAA), to NASA s Health Information Management System (10HIMS) and Human Experimental and Research Data Records (10HERD) access requirements. Additional considerations include the presence of the data in the repositories and vetting requesters for legitimacy of their use of the data. Additionally, fair access must be ensured for intramural, as well as extramural investigators. All of this must be considered in the formulation of the charters, policies and workflows for the human data repositories at NASA.

  2. MetaboLights: An Open-Access Database Repository for Metabolomics Data.

    PubMed

    Kale, Namrata S; Haug, Kenneth; Conesa, Pablo; Jayseelan, Kalaivani; Moreno, Pablo; Rocca-Serra, Philippe; Nainala, Venkata Chandrasekhar; Spicer, Rachel A; Williams, Mark; Li, Xuefei; Salek, Reza M; Griffin, Julian L; Steinbeck, Christoph

    2016-03-24

    MetaboLights is the first general purpose, open-access database repository for cross-platform and cross-species metabolomics research at the European Bioinformatics Institute (EMBL-EBI). Based upon the open-source ISA framework, MetaboLights provides Metabolomics Standard Initiative (MSI) compliant metadata and raw experimental data associated with metabolomics experiments. Users can upload their study datasets into the MetaboLights Repository. These studies are then automatically assigned a stable and unique identifier (e.g., MTBLS1) that can be used for publication reference. The MetaboLights Reference Layer associates metabolites with metabolomics studies in the archive and is extensively annotated with data fields such as structural and chemical information, NMR and MS spectra, target species, metabolic pathways, and reactions. The database is manually curated with no specific release schedules. MetaboLights is also recommended by journals for metabolomics data deposition. This unit provides a guide to using MetaboLights, downloading experimental data, and depositing metabolomics datasets using user-friendly submission tools. Copyright © 2016 John Wiley & Sons, Inc.

  3. Behavioral and Physiological Neural Network Analyses: A Common Pathway toward Pattern Recognition and Prediction

    ERIC Educational Resources Information Center

    Ninness, Chris; Lauter, Judy L.; Coffee, Michael; Clary, Logan; Kelly, Elizabeth; Rumph, Marilyn; Rumph, Robin; Kyle, Betty; Ninness, Sharon K.

    2012-01-01

    Using 3 diversified datasets, we explored the pattern-recognition ability of the Self-Organizing Map (SOM) artificial neural network as applied to diversified nonlinear data distributions in the areas of behavioral and physiological research. Experiment 1 employed a dataset obtained from the UCI Machine Learning Repository. Data for this study…

  4. Win–win data sharing in neuroscience

    PubMed Central

    Ascoli, Giorgio A; Maraver, Patricia; Nanda, Sumit; Polavaram, Sridevi; Armañanzas, Rubén

    2017-01-01

    Most neuroscientists have yet to embrace a culture of data sharing. Using our decade-long experience at NeuroMorpho.Org as an example, we discuss how publicly available repositories may benefit data producers and end-users alike. We outline practical recipes for resource developers to maximize the research impact of data sharing platforms for both contributors and users. PMID:28139675

  5. Engaging Researchers with the World's First Scholarly Arts Repositories: Ten Years after the UK's Kultur Project

    ERIC Educational Resources Information Center

    Meece, Stephanie; Robinson, Amy; Gramstadt, Marie-Therese

    2017-01-01

    Open access institutional repositories can be ill-equipped to manage the complexity of research outputs from departments of fine arts, media, drama, music, cultural heritage, and the creative arts in general. The U.K.-based Kultur project was funded to create a flexible multimedia repository model using EPrints software. The project launched the…

  6. The Emperor's New Repository

    ERIC Educational Resources Information Center

    Chudnov, Daniel

    2008-01-01

    The author does not know the first thing about building digital repositories. Maybe that is a strange thing to say, given that he works in a repository development group now, worked on the original DSpace project years ago, and worked on a few repository research projects in between. Given how long he has been around people and projects aiming to…

  7. Permanent Disposal of Nuclear Waste in Salt

    NASA Astrophysics Data System (ADS)

    Hansen, F. D.

    2016-12-01

    Salt formations hold promise for eternal removal of nuclear waste from our biosphere. Germany and the United States have ample salt formations for this purpose, ranging from flat-bedded formations to geologically mature dome structures. Both nations are revisiting nuclear waste disposal options, accompanied by extensive collaboration on applied salt repository research, design, and operation. Salt formations provide isolation while geotechnical barriers reestablish impermeability after waste is placed in the geology. Between excavation and closure, physical, mechanical, thermal, chemical, and hydrological processes ensue. Salt response over a range of stress and temperature has been characterized for decades. Research practices employ refined test techniques and controls, which improve parameter assessment for features of the constitutive models. Extraordinary computational capabilities require exacting understanding of laboratory measurements and objective interpretation of modeling results. A repository for heat-generative nuclear waste provides an engineering challenge beyond common experience. Long-term evolution of the underground setting is precluded from direct observation or measurement. Therefore, analogues and modeling predictions are necessary to establish enduring safety functions. A strong case for granular salt reconsolidation and a focused research agenda support salt repository concepts that include safety-by-design. Sandia National Laboratories is a multi-program laboratory managed and operated by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. Department of Energy's National Nuclear Security Administration under contract DE-AC04-94AL85000. Author: F. D. Hansen, Sandia National Laboratories

  8. Classifying Clinical Trial Eligibility Criteria to Facilitate Phased Cohort Identification Using Clinical Data Repositories.

    PubMed

    Wang, Amy Y; Lancaster, William J; Wyatt, Matthew C; Rasmussen, Luke V; Fort, Daniel G; Cimino, James J

    2017-01-01

    A major challenge in using electronic health record repositories for research is the difficulty matching subject eligibility criteria to query capabilities of the repositories. We propose categories for study criteria corresponding to the effort needed for querying those criteria: "easy" (supporting automated queries), mixed (initial automated querying with manual review), "hard" (fully manual record review), and "impossible" or "point of enrollment" (not typically in health repositories). We obtained a sample of 292 criteria from 20 studies from ClinicalTrials.gov. Six independent reviewers, three each from two academic research institutions, rated criteria according to our four types. We observed high interrater reliability both within and between institutions. The analysis demonstrated typical features of criteria that map with varying levels of difficulty to repositories. We propose using these features to improve enrollment workflow through more standardized study criteria, self-service repository queries, and analyst-mediated retrievals.

  9. Classifying Clinical Trial Eligibility Criteria to Facilitate Phased Cohort Identification Using Clinical Data Repositories

    PubMed Central

    Wang, Amy Y.; Lancaster, William J.; Wyatt, Matthew C.; Rasmussen, Luke V.; Fort, Daniel G.; Cimino, James J.

    2017-01-01

    A major challenge in using electronic health record repositories for research is the difficulty matching subject eligibility criteria to query capabilities of the repositories. We propose categories for study criteria corresponding to the effort needed for querying those criteria: “easy” (supporting automated queries), mixed (initial automated querying with manual review), “hard” (fully manual record review), and “impossible” or “point of enrollment” (not typically in health repositories). We obtained a sample of 292 criteria from 20 studies from ClinicalTrials.gov. Six independent reviewers, three each from two academic research institutions, rated criteria according to our four types. We observed high interrater reliability both within and between institutions. The analysis demonstrated typical features of criteria that map with varying levels of difficulty to repositories. We propose using these features to improve enrollment workflow through more standardized study criteria, self-service repository queries, and analyst-mediated retrievals. PMID:29854246

  10. Ciênsação: gaining a feeling for sciences

    NASA Astrophysics Data System (ADS)

    Abreu de Oliveira, Marcos Henrique; Fischer, Robert

    2017-03-01

    Ciênsação, an open online repository for hands-on experiments, has been developed to convince teachers in Latin America that science is best experienced first hand. Permitting students to experiment autonomously in small groups can be a challenging endeavour for educators in these countries. We analyse the reasons that cause hesitation of teachers appling hands-on experiments in class, and discuss how Ciênsação was implemented to overcome these obstacles. The resulting student research activities are specifically designed to be easily integrated in formal science education at school, to foster research skills and, most importantly, to let young people get to know science as an engaging creative activity.

  11. [Tissue repositories for research at Sheba Medical Center(SMC].

    PubMed

    Cohen, Yehudit; Barshack, Iris; Onn, Amir

    2013-06-01

    Cancer is the number one cause of death in both genders. Breakthroughs in the understanding of cancer biology, the identification of prognostic factors, and the development of new treatments are increasingly dependent on access to human cancer tissues with linked clinicopathological data. Access to human tumor samples and a large investment in translational research are needed to advance this research. The SMC tissue repositories provide researchers with biological materials, which are essential tools for cancer research. SMC tissue repositories for research aim to collect, document and preserve human biospecimens from patients with cancerous diseases. This is in order to provide the highest quality and well annotated biological biospecimens, used as essential tools to achieve the growing demands of scientific research needs. Such repositories are partners in acceLerating biomedical research and medical product development through clinical resources, in order to apply best options to the patients. Following Institutional Review Board approval and signing an Informed Consent Form, the tumor and tumor-free specimens are coLLected by a designated pathologist at the operating room only when there is a sufficient amount of the tumor, in excess of the routine needs. Blood samples are collected prior to the procedure. Other types of specimens collected include ascites fluid, pleural effusion, tissues for Optimal Cutting Temperature [OCT] and primary culture etc. Demographic, clinical, pathologicaL, and follow-up data are collected in a designated database. SMC has already established several organ or disease-specific tissue repositories within different departments. The foundation of tissue repositories requires the concentrated effort of a multidisciplinary team composed of paramedical, medical and scientific professionals. Research projects using these specimens facilitate the development of 'targeted therapy', accelerate basic research aimed at clarifying molecular mechanisms involved in cancer, and support the development of novel diagnostic tools.

  12. Deep Boreholes Seals Subjected to High P,T conditions - Proposed Experimental Studies

    NASA Astrophysics Data System (ADS)

    Caporuscio, F.

    2015-12-01

    Deep borehole experimental work will constrain the P,T conditions which "seal" material will experience in deep borehole crystalline rock repositories. The rocks of interest to this study include mafic (amphibolites) and silicic (granitic gneiss) end members. The experiments will systematically add components to capture discrete changes in both water and EBS component chemistries. Experiments in the system wall rock-clay-concrete-groundwater will evaluate interactions among components, including: mineral phase stability, metal corrosion rates and thermal limits. Based on engineered barrier studies, experimental investigations will move forward with three focusses. First, evaluation of interaction between "seal" materials and repository wall rock (crystalline) under fluid-saturated conditions over long-term (i.e., six-month) experiments; which reproduces the thermal pulse event of a repository. Second, perform experiments to determine the stability of zeolite minerals (analcime-wairakitess) under repository conditions. Both sets of experiments are critically important for understanding mineral paragenesis (zeolites and/or clay transformations) associated with "seals" in contact with wall rock at elevated temperatures. Third, mineral growth at the metal interface is a principal control on the survivability (i.e. corrosion) of waste canisters in a repository. The objective of this planned experimental work is to evaluate physio-chemical processes for 'seal' components and materials relevant to deep borehole disposal. These evaluations will encompass multi-laboratory efforts for the development of seals concepts and application of Thermal-Mechanical-Chemical (TMC) modeling work to assess barrier material interactions with subsurface fluids and other barrier materials, their stability at high temperatures, and the implications of these processes to the evaluation of thermal limits.

  13. Publishing descriptions of non-public clinical datasets: proposed guidance for researchers, repositories, editors and funding organisations.

    PubMed

    Hrynaszkiewicz, Iain; Khodiyar, Varsha; Hufton, Andrew L; Sansone, Susanna-Assunta

    2016-01-01

    Sharing of experimental clinical research data usually happens between individuals or research groups rather than via public repositories, in part due to the need to protect research participant privacy. This approach to data sharing makes it difficult to connect journal articles with their underlying datasets and is often insufficient for ensuring access to data in the long term. Voluntary data sharing services such as the Yale Open Data Access (YODA) and Clinical Study Data Request (CSDR) projects have increased accessibility to clinical datasets for secondary uses while protecting patient privacy and the legitimacy of secondary analyses but these resources are generally disconnected from journal articles-where researchers typically search for reliable information to inform future research. New scholarly journal and article types dedicated to increasing accessibility of research data have emerged in recent years and, in general, journals are developing stronger links with data repositories. There is a need for increased collaboration between journals, data repositories, researchers, funders, and voluntary data sharing services to increase the visibility and reliability of clinical research. Using the journal Scientific Data as a case study, we propose and show examples of changes to the format and peer-review process for journal articles to more robustly link them to data that are only available on request. We also propose additional features for data repositories to better accommodate non-public clinical datasets, including Data Use Agreements (DUAs).

  14. Raven-II: an open platform for surgical robotics research.

    PubMed

    Hannaford, Blake; Rosen, Jacob; Friedman, Diana W; King, Hawkeye; Roan, Phillip; Cheng, Lei; Glozman, Daniel; Ma, Ji; Kosari, Sina Nia; White, Lee

    2013-04-01

    The Raven-II is a platform for collaborative research on advances in surgical robotics. Seven universities have begun research using this platform. The Raven-II system has two 3-DOF spherical positioning mechanisms capable of attaching interchangeable four DOF instruments. The Raven-II software is based on open standards such as Linux and ROS to maximally facilitate software development. The mechanism is robust enough for repeated experiments and animal surgery experiments, but is not engineered to sufficient safety standards for human use. Mechanisms in place for interaction among the user community and dissemination of results include an electronic forum, an online software SVN repository, and meetings and workshops at major robotics conferences.

  15. HepSim: A repository with predictions for high-energy physics experiments

    DOE PAGES

    Chekanov, S. V.

    2015-02-03

    A file repository for calculations of cross sections and kinematic distributions using Monte Carlo generators for high-energy collisions is discussed. The repository is used to facilitate effective preservation and archiving of data from theoretical calculations and for comparisons with experimental data. The HepSim data library is publicly accessible and includes a number of Monte Carlo event samples with Standard Model predictions for current and future experiments. The HepSim project includes a software package to automate the process of downloading and viewing online Monte Carlo event samples. Data streaming over a network for end-user analysis is discussed.

  16. The Research Library's Role in Digital Repository Services: Final Report of the ARL Digital Repository Issues Task Force

    ERIC Educational Resources Information Center

    Association of Research Libraries, 2009

    2009-01-01

    Libraries are making diverse contributions to the development of many types of digital repositories, particularly those housing locally created digital content, including new digital objects or digitized versions of locally held works. In some instances, libraries are managing a repository and its related services entirely on their own, but often…

  17. Analysis of Academic Attitudes and Existing Processes to Inform the Design of Teaching and Learning Material Repositories: A User-Centred Approach

    ERIC Educational Resources Information Center

    King, Melanie; Loddington, Steve; Manuel, Sue; Oppenheim, Charles

    2008-01-01

    The last couple of years have brought a rise in the number of institutional repositories throughout the world and within UK Higher Education institutions, with the majority of these repositories being devoted to research output. Repositories containing teaching and learning material are less common and the workflows and business processes…

  18. The National Institutes of Health's Biomedical Translational Research Information System (BTRIS): Design, Contents, Functionality and Experience to Date

    PubMed Central

    Cimino, James J.; Ayres, Elaine J.; Remennik, Lyubov; Rath, Sachi; Freedman, Robert; Beri, Andrea; Chen, Yang; Huser, Vojtech

    2013-01-01

    The US National Institutes of Health (NIH) has developed the Biomedical Translational Research Information System (BTRIS) to support researchers’ access to translational and clinical data. BTRIS includes a data repository, a set of programs for loading data from NIH electronic health records and research data management systems, an ontology for coding the disparate data with a single terminology, and a set of user interface tools that provide access to identified data from individual research studies and data across all studies from which individually identifiable data have been removed. This paper reports on unique design elements of the system, progress to date and user experience after five years of development and operation. PMID:24262893

  19. The JRC Nanomaterials Repository: A unique facility providing representative test materials for nanoEHS research.

    PubMed

    Totaro, Sara; Cotogno, Giulio; Rasmussen, Kirsten; Pianella, Francesca; Roncaglia, Marco; Olsson, Heidi; Riego Sintes, Juan M; Crutzen, Hugues P

    2016-11-01

    The European Commission has established a Nanomaterials Repository that hosts industrially manufactured nanomaterials that are distributed world-wide for safety testing of nanomaterials. In a first instance these materials were tested in the OECD Testing Programme. They have then also been tested in several EU funded research projects. The JRC Repository of Nanomaterials has thus developed into serving the global scientific community active in the nanoEHS (regulatory) research. The unique Repository facility is a state-of-the-art installation that allows customised sub-sampling under the safest possible conditions, with traceable final sample vials distributed world-wide for research purposes. This paper describes the design of the Repository to perform a semi-automated subsampling procedure, offering high degree of flexibility and precision in the preparation of NM vials for customers, while guaranteeing the safety of the operators, and environmental protection. The JRC nanomaterials are representative for part of the world NMs market. Their wide use world-wide facilitates the generation of comparable and reliable experimental results and datasets in (regulatory) research by the scientific community, ultimately supporting the further development of the OECD regulatory test guidelines. Copyright © 2016 The Authors. Published by Elsevier Inc. All rights reserved.

  20. Connecting the pieces: Using ORCIDs to improve research impact and repositories.

    PubMed

    Baessa, Mohamed; Lery, Thibaut; Grenz, Daryl; Vijayakumar, J K

    2015-01-01

    Quantitative data are crucial in the assessment of research impact in the academic world. However, as a young university created in 2009, King Abdullah University of Science and Technology (KAUST) needs to aggregate bibliometrics from researchers coming from diverse origins, not necessarily with the proper affiliations. In this context, the University has launched an institutional repository in September 2012 with the objectives of creating a home for the intellectual outputs of KAUST researchers. Later, the university adopted the first mandated institutional open access policy in the Arab region, effective June 31, 2014. Several projects were then initiated in order to accurately identify the research being done by KAUST authors and bring it into the repository in accordance with the open access policy. Integration with ORCID has been a key element in this process and the best way to ensure data quality for researcher's scientific contributions. It included the systematic inclusion and creation, if necessary, of ORCID identifiers in the existing repository system, an institutional membership in ORCID, and the creation of dedicated integration tools. In addition and in cooperation with the Office of Research Evaluation, the Library worked at implementing a Current Research Information System (CRIS) as a standardized common resource to monitor KAUST research outputs. We will present our findings about the CRIS implementation, the ORCID API, the repository statistics as well as our approach in conducting the assessment of research impact in terms of usage by the global research community.

  1. Experiences with the BSCW Shared Workspace System as the Backbone of a Virtual Learning Environment for Students.

    ERIC Educational Resources Information Center

    Appelt, Wolfgang; Mambrey, Peter

    The GMD (German National Research Center for Information Technology) has developed the BSCW (Basic Support for Cooperative Work) Shared Workspace system within the last four years with the goal of transforming the Web from a primarily passive information repository to an active cooperation medium. The BSCW system is a Web-based groupware tool for…

  2. Raising orphans from a metadata morass: A researcher's guide to re-use of public 'omics data.

    PubMed

    Bhandary, Priyanka; Seetharam, Arun S; Arendsee, Zebulun W; Hur, Manhoi; Wurtele, Eve Syrkin

    2018-02-01

    More than 15 petabases of raw RNAseq data is now accessible through public repositories. Acquisition of other 'omics data types is expanding, though most lack a centralized archival repository. Data-reuse provides tremendous opportunity to extract new knowledge from existing experiments, and offers a unique opportunity for robust, multi-'omics analyses by merging metadata (information about experimental design, biological samples, protocols) and data from multiple experiments. We illustrate how predictive research can be accelerated by meta-analysis with a study of orphan (species-specific) genes. Computational predictions are critical to infer orphan function because their coding sequences provide very few clues. The metadata in public databases is often confusing; a test case with Zea mays mRNA seq data reveals a high proportion of missing, misleading or incomplete metadata. This metadata morass significantly diminishes the insight that can be extracted from these data. We provide tips for data submitters and users, including specific recommendations to improve metadata quality by more use of controlled vocabulary and by metadata reviews. Finally, we advocate for a unified, straightforward metadata submission and retrieval system. Copyright © 2017 Elsevier B.V. All rights reserved.

  3. Usability Evaluation of a Research Repository and Collaboration Web Site

    ERIC Educational Resources Information Center

    Zhang, Tao; Maron, Deborah J.; Charles, Christopher C.

    2013-01-01

    This article reports results from an empirical usability evaluation of Human-Animal Bond Research Initiative Central as part of the effort to develop an open access research repository and collaboration platform for human-animal bond researchers. By repurposing and altering key features of the original HUBzero system, Human-Animal Bond Research…

  4. iT2DMS: a Standard-Based Diabetic Disease Data Repository and its Pilot Experiment on Diabetic Retinopathy Phenotyping and Examination Results Integration.

    PubMed

    Wu, Huiqun; Wei, Yufang; Shang, Yujuan; Shi, Wei; Wang, Lei; Li, Jingjing; Sang, Aimin; Shi, Lili; Jiang, Kui; Dong, Jiancheng

    2018-06-06

    Type 2 diabetes mellitus (T2DM) is a common chronic disease, and the fragment data collected through separated vendors makes continuous management of DM patients difficult. The lack of standard of fragment data from those diabetic patients also makes the further potential phenotyping based on the diabetic data difficult. Traditional T2DM data repository only supports data collection from T2DM patients, lack of phenotyping ability and relied on standalone database design, limiting the secondary usage of these valuable data. To solve these issues, we proposed a novel T2DM data repository framework, which was based on standards. This repository can integrate data from various sources. It would be used as a standardized record for further data transfer as well as integration. Phenotyping was conducted based on clinical guidelines with KNIME workflow. To evaluate the phenotyping performance of the proposed system, data was collected from local community by healthcare providers and was then tested using algorithms. The results indicated that the proposed system could detect DR cases with an average accuracy of about 82.8%. Furthermore, these results had the promising potential of addressing fragmented data. The proposed system has integrating and phenotyping abilities, which could be used for diabetes research in future studies.

  5. ROSA P : The National Transportation Library’s Repository and Open Science Access Portal

    DOT National Transportation Integrated Search

    2018-01-01

    The National Transportation Library (NTL) was founded as an all-digital repository of US DOT research reports, technical publications and data products. NTLs primary public offering is ROSA P, the Repository and Open Science Access Portal. An open...

  6. Knowledge Management Systems: Linking Contribution, Refinement and Use

    ERIC Educational Resources Information Center

    Chung, Ting-ting

    2009-01-01

    Electronic knowledge repositories represent one of the fundamental tools for knowledge management (KM) initiatives. Existing research, however, has largely focused on supply-side driven research questions, such as employee motivation to contribute knowledge to a repository. This research turns attention to the dynamic relationship between the…

  7. FY16 Summary Report: Participation in the KOSINA Project

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Matteo, Edward N.; Hansen, Francis D.

    Salt formations represent a promising host for disposal of nuclear waste in the United States and Germany. Together, these countries provided fully developed safety cases for bedded salt and domal salt, respectively. Today, Germany and the United States find themselves in similar positions with respect to salt formations serving as repositories for heat-generating nuclear waste. German research centers are evaluating bedded and pillow salt formations to contrast with their previous safety case made for the Gorleben dome. Sandia National Laboratories is collaborating on this effort as an Associate Partner, and this report summarizes that teamwork. Sandia and German research groupsmore » have a long-standing cooperative approach to repository science, engineering, operations, safety assessment, testing, modeling and other elements comprising the basis for salt disposal. Germany and the United States hold annual bilateral workshops, which cover a spectrum of issues surrounding the viability of salt formations. Notably, recent efforts include development of a database for features, events, and processes applying broadly and generically to bedded and domal salt. Another international teaming activity evaluates salt constitutive models, including hundreds of new experiments conducted on bedded salt from the Waste Isolation Pilot Plant. These extensive collaborations continue to build the scientific basis for salt disposal. Repository deliberations in the United States are revisiting bedded and domal salt for housing a nuclear waste repository. By agreeing to collaborate with German peers, our nation stands to benefit by assurance of scientific position, exchange of operational concepts, and approach to elements of the safety case, all reflecting cost and time efficiency.« less

  8. Bentonite Clay Evolution at Elevated Pressures and Temperatures: An experimental study for generic nuclear repositories

    NASA Astrophysics Data System (ADS)

    Caporuscio, F. A.; Cheshire, M.; McCarney, M.

    2012-12-01

    The Used Fuel Disposition Campaign is presently engaged in looking at various generic repository options for disposal of used fuel. Of interest are the disposal of high heat load canisters ,which may allow for a reduced repository footprint. The focus of this experimental work is to characterize Engineered Barrier Systems (EBS) conditions in repositories. Clay minerals - as backfill or buffer materials - are critical to the performance of the EBS. Experiments were performed in Dickson cells at 150 bar and sequentially stepped from 125 oC to 300 oC over a period of ~1 month. An unprocessed bentonite from Colony, Wyoming was used as the buffer material in each experiment. An K-Ca-Na-Cl-rich brine (replicating deep Stripa groundwater) was used at a 9:1 water:rock ratio. The baseline experiment contained brine + clay, while three other experiments contained metals that could be used as waste form canisters (brine +clay+304SS, brine+clay+316SS, brine+clay+Cu). All experiments were buffered at the Mt-Fe oxygen fugacity univarient line. As experiment temperature increased and time progressed, pH, K and Ca ion concentrations dropped, while Si, Na, and SO4 concentrations increased. Silicon was liberated into the fluid phase (>1000 ppm) and precipitated during the quenching of the experiment. The precipitated silica transformed to cristobalite as cooling progressed. Potassium was mobilized and exchanged with interlayer Na, transitioning the clay from Na-montmorillonite to K-smectite. Though illitization was not observed in these experiments, its formation may be kinetically limited and longer-term experiments are underway to evaluate the equilibrium point in this reaction. Clinoptilolite present in the starting bentonite mixture is unstable above 150 oC. Hence, the zeolite broke down at high temperatures but recrystallized as the quench event occurred. This was borne out in SEM images that showed clinoptilolite as a very late stage growth mineral. Both experimental runs containing steel exhibit the generation of a chlorite / Fe-saponite layer at the clay-metal boundary. The formation of minor amounts of pentlandite [(Fe,Ni)9S8] also occurs on both steel plates. Chalcocite (Cu2S) formed as a corrosion product on the Cu plates. The two sulfide phases have been produced by the generation of H2S gas during the experimental runs. The H2S is formed by the breakdown of pyrite framboids at high temperature in the bentonite. Such experiments on representative EBS materials at elevated P,T repository conditions are providing useful information for generic repository studies. Lack of illite formation is common in clay experiments and may be related to kinetics or K concentration. Precipitated SiO2 may potentially seal heating cracks in the clay backfill. The chlorite layer generated on steel may act as a passivation material and prevent corrosion of the steel canister wall. Finally, even if zeolites break down during the high temperature thermal pulse of a repository, zeolites may form again as the repository inventory cools off and perform as radionuclide sorbing phases.

  9. Visualizing research collections in the National Transportation Library's digital repository : ROSA P.

    DOT National Transportation Integrated Search

    2017-01-01

    The National Transportation Library's (NTL) Repository and Open Science Portal (ROSA P) : is a digital library for transportation, including U. S. Department of Transportation : sponsored research results and technical publications, other documents a...

  10. Computational knowledge integration in biopharmaceutical research.

    PubMed

    Ficenec, David; Osborne, Mark; Pradines, Joel; Richards, Dan; Felciano, Ramon; Cho, Raymond J; Chen, Richard O; Liefeld, Ted; Owen, James; Ruttenberg, Alan; Reich, Christian; Horvath, Joseph; Clark, Tim

    2003-09-01

    An initiative to increase biopharmaceutical research productivity by capturing, sharing and computationally integrating proprietary scientific discoveries with public knowledge is described. This initiative involves both organisational process change and multiple interoperating software systems. The software components rely on mutually supporting integration techniques. These include a richly structured ontology, statistical analysis of experimental data against stored conclusions, natural language processing of public literature, secure document repositories with lightweight metadata, web services integration, enterprise web portals and relational databases. This approach has already begun to increase scientific productivity in our enterprise by creating an organisational memory (OM) of internal research findings, accessible on the web. Through bringing together these components it has also been possible to construct a very large and expanding repository of biological pathway information linked to this repository of findings which is extremely useful in analysis of DNA microarray data. This repository, in turn, enables our research paradigm to be shifted towards more comprehensive systems-based understandings of drug action.

  11. Life Sciences Data Archives (LSDA) in the Post-Shuttle Era

    NASA Technical Reports Server (NTRS)

    Fitts, Mary A.; Johnson-Throop, Kathy; Havelka, Jacque; Thomas, Diedre

    2010-01-01

    Now, more than ever before, NASA is realizing the value and importance of their intellectual assets. Principles of knowledge management-the systematic use and reuse of information, experience, and expertise to achieve a specific goal-are being applied throughout the agency. LSDA is also applying these solutions, which rely on a combination of content and collaboration technologies, to enable research teams to create, capture, share, and harness knowledge to do the things they do well, even better. In the early days of spaceflight, space life sciences data were collected and stored in numerous databases, formats, media-types and geographical locations. These data were largely unknown/unavailable to the research community. The Biomedical Informatics and Health Care Systems Branch of the Space Life Sciences Directorate at JSC and the Data Archive Project at ARC, with funding from the Human Research Program through the Exploration Medical Capability Element, are fulfilling these requirements through the systematic population of the Life Sciences Data Archive. This project constitutes a formal system for the acquisition, archival and distribution of data for HRP-related experiments and investigations. The general goal of the archive is to acquire, preserve, and distribute these data and be responsive to inquiries for the science communities. Information about experiments and data, as well as non-attributable human data and data from other species' are available on our public Web site http://lsda.jsc.nasa.gov. The Web site also includes a repository for biospecimens, and a utilization process. NASA has undertaken an initiative to develop a Shuttle Data Archive repository. The Shuttle program is nearing its end in 2010 and it is critical that the medical and research data related to the Shuttle program be captured, retained, and usable for research, lessons learned, and future mission planning. Communities of practice are groups of people who share a concern or a passion for something they do, and learn how to do it better as they interact regularly. LSDA works with the HRP community of practice to ensure that we are preserving the relevant research and data they need in the LSDA repository. An evidence-based approach to risk management is required in space life sciences. Evidence changes over time. LSDA has a pilot project with Collexis, a new type of Web-based search engine. Collexis differentiates itself from full-text search engines by making use of thesauri for information retrieval. The high-quality search is based on semantics that have been defined in a life sciences ontology. Additionally, Collexis' matching technology is unique, allowing discovery of partially matching dicuments. Users do not have to construct a complicated (Boolean) search query, but can simply enter a free text search without the risk of getting "no results". Collexis may address these issues by virtue of its retrieval and discovery capabilities across multiple repositories.

  12. Doing Your Science While You're in Orbit

    NASA Astrophysics Data System (ADS)

    Green, Mark L.; Miller, Stephen D.; Vazhkudai, Sudharshan S.; Trater, James R.

    2010-11-01

    Large-scale neutron facilities such as the Spallation Neutron Source (SNS) located at Oak Ridge National Laboratory need easy-to-use access to Department of Energy Leadership Computing Facilities and experiment repository data. The Orbiter thick- and thin-client and its supporting Service Oriented Architecture (SOA) based services (available at https://orbiter.sns.gov) consist of standards-based components that are reusable and extensible for accessing high performance computing, data and computational grid infrastructure, and cluster-based resources easily from a user configurable interface. The primary Orbiter system goals consist of (1) developing infrastructure for the creation and automation of virtual instrumentation experiment optimization, (2) developing user interfaces for thin- and thick-client access, (3) provide a prototype incorporating major instrument simulation packages, and (4) facilitate neutron science community access and collaboration. The secure Orbiter SOA authentication and authorization is achieved through the developed Virtual File System (VFS) services, which use Role-Based Access Control (RBAC) for data repository file access, thin-and thick-client functionality and application access, and computational job workflow management. The VFS Relational Database Management System (RDMS) consists of approximately 45 database tables describing 498 user accounts with 495 groups over 432,000 directories with 904,077 repository files. Over 59 million NeXus file metadata records are associated to the 12,800 unique NeXus file field/class names generated from the 52,824 repository NeXus files. Services that enable (a) summary dashboards of data repository status with Quality of Service (QoS) metrics, (b) data repository NeXus file field/class name full text search capabilities within a Google like interface, (c) fully functional RBAC browser for the read-only data repository and shared areas, (d) user/group defined and shared metadata for data repository files, (e) user, group, repository, and web 2.0 based global positioning with additional service capabilities are currently available. The SNS based Orbiter SOA integration progress with the Distributed Data Analysis for Neutron Scattering Experiments (DANSE) software development project is summarized with an emphasis on DANSE Central Services and the Virtual Neutron Facility (VNF). Additionally, the DANSE utilization of the Orbiter SOA authentication, authorization, and data transfer services best practice implementations are presented.

  13. NCBI GEO: archive for high-throughput functional genomic data.

    PubMed

    Barrett, Tanya; Troup, Dennis B; Wilhite, Stephen E; Ledoux, Pierre; Rudnev, Dmitry; Evangelista, Carlos; Kim, Irene F; Soboleva, Alexandra; Tomashevsky, Maxim; Marshall, Kimberly A; Phillippy, Katherine H; Sherman, Patti M; Muertter, Rolf N; Edgar, Ron

    2009-01-01

    The Gene Expression Omnibus (GEO) at the National Center for Biotechnology Information (NCBI) is the largest public repository for high-throughput gene expression data. Additionally, GEO hosts other categories of high-throughput functional genomic data, including those that examine genome copy number variations, chromatin structure, methylation status and transcription factor binding. These data are generated by the research community using high-throughput technologies like microarrays and, more recently, next-generation sequencing. The database has a flexible infrastructure that can capture fully annotated raw and processed data, enabling compliance with major community-derived scientific reporting standards such as 'Minimum Information About a Microarray Experiment' (MIAME). In addition to serving as a centralized data storage hub, GEO offers many tools and features that allow users to effectively explore, analyze and download expression data from both gene-centric and experiment-centric perspectives. This article summarizes the GEO repository structure, content and operating procedures, as well as recently introduced data mining features. GEO is freely accessible at http://www.ncbi.nlm.nih.gov/geo/.

  14. DataUp: Helping manage and archive data within the researcher's workflow

    NASA Astrophysics Data System (ADS)

    Strasser, C.

    2012-12-01

    There are many barriers to data management and sharing among earth and environmental scientists; among the most significant are lacks of knowledge about best practices for data management, metadata standards, or appropriate data repositories for archiving and sharing data. We have developed an open-source add-in for Excel and an open source web application intended to help researchers overcome these barriers. DataUp helps scientists to (1) determine whether their file is CSV compatible, (2) generate metadata in a standard format, (3) retrieve an identifier to facilitate data citation, and (4) deposit their data into a repository. The researcher does not need a prior relationship with a data repository to use DataUp; the newly implemented ONEShare repository, a DataONE member node, is available for any researcher to archive and share their data. By meeting researchers where they already work, in spreadsheets, DataUp becomes part of the researcher's workflow and data management and sharing becomes easier. Future enhancement of DataUp will rely on members of the community adopting and adapting the DataUp tools to meet their unique needs, including connecting to analytical tools, adding new metadata schema, and expanding the list of connected data repositories. DataUp is a collaborative project between Microsoft Research Connections, the University of California's California Digital Library, the Gordon and Betty Moore Foundation, and DataONE.

  15. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Leigh, Christi D.; Hansen, Francis D.

    This report summarizes the state of salt repository science, reviews many of the technical issues pertaining to disposal of heat-generating nuclear waste in salt, and proposes several avenues for future science-based activities to further the technical basis for disposal in salt. There are extensive salt formations in the forty-eight contiguous states, and many of them may be worthy of consideration for nuclear waste disposal. The United States has extensive experience in salt repository sciences, including an operating facility for disposal of transuranic wastes. The scientific background for salt disposal including laboratory and field tests at ambient and elevated temperature, principlesmore » of salt behavior, potential for fracture damage and its mitigation, seal systems, chemical conditions, advanced modeling capabilities and near-future developments, performance assessment processes, and international collaboration are all discussed. The discussion of salt disposal issues is brought current, including a summary of recent international workshops dedicated to high-level waste disposal in salt. Lessons learned from Sandia National Laboratories' experience on the Waste Isolation Pilot Plant and the Yucca Mountain Project as well as related salt experience with the Strategic Petroleum Reserve are applied in this assessment. Disposal of heat-generating nuclear waste in a suitable salt formation is attractive because the material is essentially impermeable, self-sealing, and thermally conductive. Conditions are chemically beneficial, and a significant experience base exists in understanding this environment. Within the period of institutional control, overburden pressure will seal fractures and provide a repository setting that limits radionuclide movement. A salt repository could potentially achieve total containment, with no releases to the environment in undisturbed scenarios for as long as the region is geologically stable. Much of the experience gained from United States repository development, such as seal system design, coupled process simulation, and application of performance assessment methodology, helps define a clear strategy for a heat-generating nuclear waste repository in salt.« less

  16. Organizing Scientific Data Sets: Studying Similarities and Differences in Metadata and Subject Term Creation

    ERIC Educational Resources Information Center

    White, Hollie C.

    2012-01-01

    Background: According to Salo (2010), the metadata entered into repositories are "disorganized" and metadata schemes underlying repositories are "arcane". This creates a challenging repository environment in regards to personal information management (PIM) and knowledge organization systems (KOSs). This dissertation research is…

  17. TRAC Searchable Research Library

    DTIC Science & Technology

    2016-05-01

    network accessible document repository for technical documents and similar document artifacts. We used a model-based approach using the Vector...demonstration and model refinement. 14. SUBJECT TERMS Knowledge Management, Document Repository , Digital Library, Vector Directional Data Model...27 Figure D1. Administrator Repository Upload Page. ................................................................... D-2 Figure D2

  18. Biomedical Data Sharing and Reuse: Attitudes and Practices of Clinical and Scientific Research Staff.

    PubMed

    Federer, Lisa M; Lu, Ya-Ling; Joubert, Douglas J; Welsh, Judith; Brandys, Barbara

    2015-01-01

    Significant efforts are underway within the biomedical research community to encourage sharing and reuse of research data in order to enhance research reproducibility and enable scientific discovery. While some technological challenges do exist, many of the barriers to sharing and reuse are social in nature, arising from researchers' concerns about and attitudes toward sharing their data. In addition, clinical and basic science researchers face their own unique sets of challenges to sharing data within their communities. This study investigates these differences in experiences with and perceptions about sharing data, as well as barriers to sharing among clinical and basic science researchers. Clinical and basic science researchers in the Intramural Research Program at the National Institutes of Health were surveyed about their attitudes toward and experiences with sharing and reusing research data. Of 190 respondents to the survey, the 135 respondents who identified themselves as clinical or basic science researchers were included in this analysis. Odds ratio and Fisher's exact tests were the primary methods to examine potential relationships between variables. Worst-case scenario sensitivity tests were conducted when necessary. While most respondents considered data sharing and reuse important to their work, they generally rated their expertise as low. Sharing data directly with other researchers was common, but most respondents did not have experience with uploading data to a repository. A number of significant differences exist between the attitudes and practices of clinical and basic science researchers, including their motivations for sharing, their reasons for not sharing, and the amount of work required to prepare their data. Even within the scope of biomedical research, addressing the unique concerns of diverse research communities is important to encouraging researchers to share and reuse data. Efforts at promoting data sharing and reuse should be aimed at solving not only technological problems, but also addressing researchers' concerns about sharing their data. Given the varied practices of individual researchers and research communities, standardizing data practices like data citation and repository upload could make sharing and reuse easier.

  19. Practical management of heterogeneous neuroimaging metadata by global neuroimaging data repositories

    PubMed Central

    Neu, Scott C.; Crawford, Karen L.; Toga, Arthur W.

    2012-01-01

    Rapidly evolving neuroimaging techniques are producing unprecedented quantities of digital data at the same time that many research studies are evolving into global, multi-disciplinary collaborations between geographically distributed scientists. While networked computers have made it almost trivial to transmit data across long distances, collecting and analyzing this data requires extensive metadata if the data is to be maximally shared. Though it is typically straightforward to encode text and numerical values into files and send content between different locations, it is often difficult to attach context and implicit assumptions to the content. As the number of and geographic separation between data contributors grows to national and global scales, the heterogeneity of the collected metadata increases and conformance to a single standardization becomes implausible. Neuroimaging data repositories must then not only accumulate data but must also consolidate disparate metadata into an integrated view. In this article, using specific examples from our experiences, we demonstrate how standardization alone cannot achieve full integration of neuroimaging data from multiple heterogeneous sources and why a fundamental change in the architecture of neuroimaging data repositories is needed instead. PMID:22470336

  20. Practical management of heterogeneous neuroimaging metadata by global neuroimaging data repositories.

    PubMed

    Neu, Scott C; Crawford, Karen L; Toga, Arthur W

    2012-01-01

    Rapidly evolving neuroimaging techniques are producing unprecedented quantities of digital data at the same time that many research studies are evolving into global, multi-disciplinary collaborations between geographically distributed scientists. While networked computers have made it almost trivial to transmit data across long distances, collecting and analyzing this data requires extensive metadata if the data is to be maximally shared. Though it is typically straightforward to encode text and numerical values into files and send content between different locations, it is often difficult to attach context and implicit assumptions to the content. As the number of and geographic separation between data contributors grows to national and global scales, the heterogeneity of the collected metadata increases and conformance to a single standardization becomes implausible. Neuroimaging data repositories must then not only accumulate data but must also consolidate disparate metadata into an integrated view. In this article, using specific examples from our experiences, we demonstrate how standardization alone cannot achieve full integration of neuroimaging data from multiple heterogeneous sources and why a fundamental change in the architecture of neuroimaging data repositories is needed instead.

  1. Research on wild relatives of fruit and nut crops at the Davis repository

    USDA-ARS?s Scientific Manuscript database

    The USDA germplasm repository in Davis is responsible for acquiring, conserving and distributing a broad spectrum of diversity of subtropical and temperate fruit and nut species germplasm to stakeholders around the world. Currently the repository holds over 7000 accessions of germplasm including Act...

  2. Building Connections, Collections, and Communities: Increasing the Visibility and Impact of Extension through Institutional Repositories

    ERIC Educational Resources Information Center

    Inefuku, Harrison W.; Franz, Nancy K.

    2015-01-01

    Over the past 20 years, university libraries have developed and manage institutional repositories--digital libraries that provide free, public access to the research, scholarship, and publications of their university's faculty, staff, and students. Although underused by Extension professionals, institutional repositories are powerful tools that…

  3. Embracing the Future: Embedding Digital Repositories in Higher Education Institutions. Research Brief

    ERIC Educational Resources Information Center

    Hoorens, Stijn; van Dijk, Lidia Villalba; van Stolk, Christian

    2009-01-01

    This briefing paper captures the key findings and recommendations of a study commissioned by the Joint Information Systems Committee on aspects of the strategic commitment of institutions to repository sustainability. This project, labelled EMBRACE (EMBedding Repositories And Consortial Enhancement), is aimed at enhancing the functionality,…

  4. The open research system: a web-based metadata and data repository for collaborative research

    Treesearch

    Charles M. Schweik; Alexander Stepanov; J. Morgan Grove

    2005-01-01

    Beginning in 1999, a web-based metadata and data repository we call the "open research system" (ORS) was designed and built to assist geographically distributed scientific research teams. The purpose of this innovation was to promote the open sharing of data within and across organizational lines and across geographic distances. As the use of the system...

  5. Business Process Flow Diagrams in Tissue Bank Informatics System Design, and Identification and Communication of Best Practices: The Pharmaceutical Industry Experience.

    PubMed

    McDonald, Sandra A; Velasco, Elizabeth; Ilasi, Nicholas T

    2010-12-01

    Pfizer, Inc.'s Tissue Bank, in conjunction with Pfizer's BioBank (biofluid repository), endeavored to create an overarching internal software package to cover all general functions of both research facilities, including sample receipt, reconciliation, processing, storage, and ordering. Business process flow diagrams were developed by the Tissue Bank and Informatics teams as a way of characterizing best practices both within the Bank and in its interactions with key internal and external stakeholders. Besides serving as a first step for the software development, such formalized process maps greatly assisted the identification and communication of best practices and the optimization of current procedures. The diagrams shared here could assist other biospecimen research repositories (both pharmaceutical and other settings) for comparative purposes or as a guide to successful informatics design. Therefore, it is recommended that biorepositories consider establishing formalized business process flow diagrams for their laboratories, to address these objectives of communication and strategy.

  6. FAIRDOMHub: a repository and collaboration environment for sharing systems biology research.

    PubMed

    Wolstencroft, Katherine; Krebs, Olga; Snoep, Jacky L; Stanford, Natalie J; Bacall, Finn; Golebiewski, Martin; Kuzyakiv, Rostyk; Nguyen, Quyen; Owen, Stuart; Soiland-Reyes, Stian; Straszewski, Jakub; van Niekerk, David D; Williams, Alan R; Malmström, Lars; Rinn, Bernd; Müller, Wolfgang; Goble, Carole

    2017-01-04

    The FAIRDOMHub is a repository for publishing FAIR (Findable, Accessible, Interoperable and Reusable) Data, Operating procedures and Models (https://fairdomhub.org/) for the Systems Biology community. It is a web-accessible repository for storing and sharing systems biology research assets. It enables researchers to organize, share and publish data, models and protocols, interlink them in the context of the systems biology investigations that produced them, and to interrogate them via API interfaces. By using the FAIRDOMHub, researchers can achieve more effective exchange with geographically distributed collaborators during projects, ensure results are sustained and preserved and generate reproducible publications that adhere to the FAIR guiding principles of data stewardship. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  7. Accessing and integrating data and knowledge for biomedical research.

    PubMed

    Burgun, A; Bodenreider, O

    2008-01-01

    To review the issues that have arisen with the advent of translational research in terms of integration of data and knowledge, and survey current efforts to address these issues. Using examples form the biomedical literature, we identified new trends in biomedical research and their impact on bioinformatics. We analyzed the requirements for effective knowledge repositories and studied issues in the integration of biomedical knowledge. New diagnostic and therapeutic approaches based on gene expression patterns have brought about new issues in the statistical analysis of data, and new workflows are needed are needed to support translational research. Interoperable data repositories based on standard annotations, infrastructures and services are needed to support the pooling and meta-analysis of data, as well as their comparison to earlier experiments. High-quality, integrated ontologies and knowledge bases serve as a source of prior knowledge used in combination with traditional data mining techniques and contribute to the development of more effective data analysis strategies. As biomedical research evolves from traditional clinical and biological investigations towards omics sciences and translational research, specific needs have emerged, including integrating data collected in research studies with patient clinical data, linking omics knowledge with medical knowledge, modeling the molecular basis of diseases, and developing tools that support in-depth analysis of research data. As such, translational research illustrates the need to bridge the gap between bioinformatics and medical informatics, and opens new avenues for biomedical informatics research.

  8. NCI Expands Repository of Cancer Research Models

    Cancer.gov

    NCI is expanding its Patient-Derived Models Repository (PDMR), which generates and distributes models like patient-derived xenografts and organoids. In this Cancer Currents Q&A with Drs. Yvonne Evrard and James Doroshow, learn how the expansion can help cancer researchers make more rapid progress.

  9. Examination of Data Accession at the National Snow and Ice Data Center

    NASA Astrophysics Data System (ADS)

    Scott, D. J.; Booker, L.

    2017-12-01

    The National Snow and Ice Data Center (NSIDC) stewards nearly 750 publicly available snow and ice data sets that support research into our world's frozen realms. NSIDC data management is primarily supported by the National Aeronautics and Space Administration (NASA), the National Science Foundation (NSF) and the National Oceanic and Atmospheric Administration (NOAA), and most of the data we archive and distribute is assigned to NSIDC through the funding agency programs. In addition to these mandates, NSIDC has historically offered data stewardship to researchers wanting to properly preserve and increase visibility of their research data under our primary programs (NASA, NSF, NOAA). With publishers now requiring researchers to deliver data to a repository prior to the publication of their data-related papers, we have seen an increase in researcher-initiated data accession requests. This increase is pushing us to reexamine our process to ensure timeliness in the acquisition and release of these data. In this presentation, we will discuss the support and value a researcher receives by submitting data to a trustworthy repository. We will examine NSIDC's data accession practices, and the challenges of a consistent process across NSIDC's multiple funding sponsors. Finally, we will share recent activities related to improving our process and ideas we have for enhancing the overall data accession experience.

  10. Digital Libraries and Repositories in India: An Evaluative Study

    ERIC Educational Resources Information Center

    Mittal, Rekha; Mahesh, G.

    2008-01-01

    Purpose: The purpose of this research is to identify and evaluate the collections within digital libraries and repositories in India available in the public domain. Design/methodology/approach: The digital libraries and repositories were identified through a study of the literature, as well as internet searching and browsing. The resulting digital…

  11. Relevant Repositories of Public Knowledge? Libraries, Museums and Archives in "The Information Age"

    ERIC Educational Resources Information Center

    Usherwood, Bob; Wilson, Kerry; Bryson, Jared

    2005-01-01

    In a project funded by the AHRB, researchers at the University of Sheffield used a combination of quantitative and qualitative research methods to examine the perceived contemporary relevance of archives, libraries and museums. The research sought to discern how far the British people value access to these established repositories of public…

  12. Researcher-library collaborations: Data repositories as a service for researchers.

    PubMed

    Gordon, Andrew S; Millman, David S; Steiger, Lisa; Adolph, Karen E; Gilmore, Rick O

    New interest has arisen in organizing, preserving, and sharing the raw materials-the data and metadata-that undergird the published products of research. Library and information scientists have valuable expertise to bring to bear in the effort to create larger, more diverse, and more widely used data repositories. However, for libraries to be maximally successful in providing the research data management and preservation services required of a successful data repository, librarians must work closely with researchers and learn about their data management workflows. Databrary is a data repository that is closely linked to the needs of a specific scholarly community-researchers who use video as a main source of data to study child development and learning. The project's success to date is a result of its focus on community outreach and providing services for scholarly communication, engaging institutional partners, offering services for data curation with the guidance of closely involved information professionals, and the creation of a strong technical infrastructure. Databrary plans to improve its curation tools that allow researchers to deposit their own data, enhance the user-facing feature set, increase integration with library systems, and implement strategies for long-term sustainability.

  13. Benefits of International Collaboration on the International Space Station

    NASA Technical Reports Server (NTRS)

    Hasbrook, Pete; Robinson, Julie A.; Brown Tate, Judy; Thumm, Tracy; Cohen, Luchino; Marcil, Isabelle; De Parolis, Lina; Hatton, Jason; Umezawa, Kazuo; Shirakawa, Masaki; hide

    2017-01-01

    The International Space Station is a valuable platform for research in space, but the benefits are limited if research is only conducted by individual countries. Through the efforts of the ISS Program Science Forum, international science working groups, and interagency cooperation, international collaboration on the ISS has expanded as ISS utilization has matured. Members of science teams benefit from working with counterparts in other countries. Scientists and institutions bring years of experience and specialized expertise to collaborative investigations, leading to new perspectives and approaches to scientific challenges. Combining new ideas and historical results brings synergy and improved peer-reviewed scientific methods and results. World-class research facilities can be expensive and logistically complicated, jeopardizing their full utilization. Experiments that would be prohibitively expensive for a single country can be achieved through contributions of resources from two or more countries, such as crew time, up- and downmass, and experiment hardware. Cooperation also avoids duplication of experiments and hardware among agencies. Biomedical experiments can be completed earlier if astronauts or cosmonauts from multiple agencies participate. Countries responding to natural disasters benefit from ISS imagery assets, even if the country has no space agency of its own. Students around the world participate in ISS educational opportunities, and work with students in other countries, through open curriculum packages and through international competitions. Even experiments conducted by a single country can benefit scientists around the world, through specimen sharing programs and publicly accessible "open data" repositories. For ISS data, these repositories include GeneLab and the Physical Science Informatics System. Scientists can conduct new research using ISS data without having to launch and execute their own experiments. Multilateral collections of research results publications, maintained by the ISS international partnership and accessible via nasa.gov, make ISS results available worldwide, and encourage new users, ideas and research. The paper explores international collaboration history, its evolution and maturation, change of focus during its different phases, and growth of its effectiveness (in accordance with the especially established criteria) in the light of benefits for the entire ISS community. With the International Space Station extended through at least 2024, more crew time becoming available and new facilities arriving on board the ISS, these benefits of international scientific collaboration on the ISS can only increase.

  14. Canary: An NLP Platform for Clinicians and Researchers.

    PubMed

    Malmasi, Shervin; Sandor, Nicolae L; Hosomura, Naoshi; Goldberg, Matt; Skentzos, Stephen; Turchin, Alexander

    2017-05-03

    Information Extraction methods can help discover critical knowledge buried in the vast repositories of unstructured clinical data. However, these methods are underutilized in clinical research, potentially due to the absence of free software geared towards clinicians with little technical expertise. The skills required for developing/using such software constitute a major barrier for medical researchers wishing to employ these methods. To address this, we have developed Canary, a free and open-source solution designed for users without natural language processing (NLP) or software engineering experience. It was designed to be fast and work out of the box via a user-friendly graphical interface.

  15. Seven [Data] Habits of Highly Successful Researchers

    NASA Astrophysics Data System (ADS)

    Kinkade, D.; Shepherd, A.; Saito, M. A.; Wiebe, P. H.; Ake, H.; Biddle, M.; Copley, N. J.; Rauch, S.; Switzer, M. E.; York, A.

    2017-12-01

    Navigating the landscape of open science and data sharing can be daunting for the long-tail scientist. From satisfying funder requirements, and ensuring proper attribution for their work, to determining the best repository for data management and archive, there are several facets to be considered. Yet, there is no single source of guidance for investigators who may be using multiple research funding models. What role can existing repositories play to help facilitate a more effective data sharing workflow? The Biological and Chemical Oceanographic Data Management Office (BCO-DMO) is a domain-specific repository occupying the niche between funder and investigator. The office works closely with its stakeholders to develop and provide guidance, services, and tools that assist researchers in meeting their data sharing needs. From determining if BCO-DMO is the appropriate repository to manage an investigator's project data, to ensuring that investigator is able to fulfill funder requirements. The goal is to relieve the investigator of the more difficult aspects of data management and data sharing, while simultaneously educating them in better data management practices that will streamline the process of conducting open research in the future. This presentation will provide an overview of the BCO-DMO repository, highlighting some of the services and guidance the office provides to its community.

  16. Creating the Evidence through Comparative Effectiveness Research for Interprofessional Education and Collaborative Practice by Deploying a National Intervention Network and a National Data Repository

    PubMed Central

    Pechacek, Judith; Cerra, Frank; Brandt, Barbara; Lutfiyya, May Nawal; Delaney, Connie

    2015-01-01

    Background: There is currently a resurgence of interest in interprofessional education and collaborative practice (IPECP) and its potential to positively impact health outcomes at both the patient level and population level, healthcare delivery, and health professions education. This resurgence of interest led to the creation of the National Center on Interprofessional Collaborative Practice and Education in October 2012. Methods: This paper describes three intertwined knowledge generation strategies of the National Center on Interprofessional Practice and Education: (1) the development of a Nexus Incubator Network, (2) the undertaking of comparative effectiveness research, and (3) the creation of a National Center Data Repository. Results: As these strategies are implemented over time they will result in the production of empirically grounded knowledge regarding the direction and scope of the impact, if any, of IPECP on well-defined health and healthcare outcomes including the possible improvement of the patient experience of care. Conclusions: Among the motivating factors for the National Center and the three strategies adopted and addressed herein is the need for rigorously produced, scientifically sound evidence regarding IPECP and whether or not it has the capacity to positively affect the patient experience of care, the health of populations, and the per capita cost of healthcare. PMID:27417753

  17. A web based tool for storing and visualising data generated within a smart home.

    PubMed

    McDonald, H A; Nugent, C D; Moore, G; Finlay, D D; Hallberg, J

    2011-01-01

    There is a growing need to re-assess the current approaches available to researchers for storing and managing heterogeneous data generated within a smart home environment. In our current work we have developed the homeML Application; a web based tool to support researchers engaged in the area of smart home research as they perform experiments. Within this paper the homeML Application is presented which includes the fundamental components of the homeML Repository and the homeML Toolkit. Results from a usability study conducted by 10 computer science researchers are presented; the initial results of which have been positive.

  18. Repository Collection Policies: Is a Liberal and Inclusive Policy Helpful or Harmful?

    ERIC Educational Resources Information Center

    Bankier, Jean-Gabriel; Smith, Courtney

    2010-01-01

    Institutional repositories have been established by most university libraries but their level of success has varied. Determining what belongs in such repositories has been the subject of some discussion but research would suggest that a broader rather than narrower compass is a positive approach to adopt. By seeking out a variety of content types,…

  19. Assessment of Self-Archiving in Institutional Repositories: Across Disciplines

    ERIC Educational Resources Information Center

    Xia, Jingfeng

    2007-01-01

    This research examined self-archiving practices by four disciplines in seven institutional repositories. By checking each individual item for its metadata and deposition status, the research found that a disciplinary culture is not obviously presented. Rather, self-archiving is regulated by a liaison system and a mandate policy.

  20. Facilitating Cohort Discovery by Enhancing Ontology Exploration, Query Management and Query Sharing for Large Clinical Data Repositories.

    PubMed

    Tao, Shiqiang; Cui, Licong; Wu, Xi; Zhang, Guo-Qiang

    2017-01-01

    To help researchers better access clinical data, we developed a prototype query engine called DataSphere for exploring large-scale integrated clinical data repositories. DataSphere expedites data importing using a NoSQL data management system and dynamically renders its user interface for concept-based querying tasks. DataSphere provides an interactive query-building interface together with query translation and optimization strategies, which enable users to build and execute queries effectively and efficiently. We successfully loaded a dataset of one million patients for University of Kentucky (UK) Healthcare into DataSphere with more than 300 million clinical data records. We evaluated DataSphere by comparing it with an instance of i2b2 deployed at UK Healthcare, demonstrating that DataSphere provides enhanced user experience for both query building and execution.

  1. Facilitating Cohort Discovery by Enhancing Ontology Exploration, Query Management and Query Sharing for Large Clinical Data Repositories

    PubMed Central

    Tao, Shiqiang; Cui, Licong; Wu, Xi; Zhang, Guo-Qiang

    2017-01-01

    To help researchers better access clinical data, we developed a prototype query engine called DataSphere for exploring large-scale integrated clinical data repositories. DataSphere expedites data importing using a NoSQL data management system and dynamically renders its user interface for concept-based querying tasks. DataSphere provides an interactive query-building interface together with query translation and optimization strategies, which enable users to build and execute queries effectively and efficiently. We successfully loaded a dataset of one million patients for University of Kentucky (UK) Healthcare into DataSphere with more than 300 million clinical data records. We evaluated DataSphere by comparing it with an instance of i2b2 deployed at UK Healthcare, demonstrating that DataSphere provides enhanced user experience for both query building and execution. PMID:29854239

  2. A repository based on a dynamically extensible data model supporting multidisciplinary research in neuroscience.

    PubMed

    Corradi, Luca; Porro, Ivan; Schenone, Andrea; Momeni, Parastoo; Ferrari, Raffaele; Nobili, Flavio; Ferrara, Michela; Arnulfo, Gabriele; Fato, Marco M

    2012-10-08

    Robust, extensible and distributed databases integrating clinical, imaging and molecular data represent a substantial challenge for modern neuroscience. It is even more difficult to provide extensible software environments able to effectively target the rapidly changing data requirements and structures of research experiments. There is an increasing request from the neuroscience community for software tools addressing technical challenges about: (i) supporting researchers in the medical field to carry out data analysis using integrated bioinformatics services and tools; (ii) handling multimodal/multiscale data and metadata, enabling the injection of several different data types according to structured schemas; (iii) providing high extensibility, in order to address different requirements deriving from a large variety of applications simply through a user runtime configuration. A dynamically extensible data structure supporting collaborative multidisciplinary research projects in neuroscience has been defined and implemented. We have considered extensibility issues from two different points of view. First, the improvement of data flexibility has been taken into account. This has been done through the development of a methodology for the dynamic creation and use of data types and related metadata, based on the definition of "meta" data model. This way, users are not constrainted to a set of predefined data and the model can be easily extensible and applicable to different contexts. Second, users have been enabled to easily customize and extend the experimental procedures in order to track each step of acquisition or analysis. This has been achieved through a process-event data structure, a multipurpose taxonomic schema composed by two generic main objects: events and processes. Then, a repository has been built based on such data model and structure, and deployed on distributed resources thanks to a Grid-based approach. Finally, data integration aspects have been addressed by providing the repository application with an efficient dynamic interface designed to enable the user to both easily query the data depending on defined datatypes and view all the data of every patient in an integrated and simple way. The results of our work have been twofold. First, a dynamically extensible data model has been implemented and tested based on a "meta" data-model enabling users to define their own data types independently from the application context. This data model has allowed users to dynamically include additional data types without the need of rebuilding the underlying database. Then a complex process-event data structure has been built, based on this data model, describing patient-centered diagnostic processes and merging information from data and metadata. Second, a repository implementing such a data structure has been deployed on a distributed Data Grid in order to provide scalability both in terms of data input and data storage and to exploit distributed data and computational approaches in order to share resources more efficiently. Moreover, data managing has been made possible through a friendly web interface. The driving principle of not being forced to preconfigured data types has been satisfied. It is up to users to dynamically configure the data model for the given experiment or data acquisition program, thus making it potentially suitable for customized applications. Based on such repository, data managing has been made possible through a friendly web interface. The driving principle of not being forced to preconfigured data types has been satisfied. It is up to users to dynamically configure the data model for the given experiment or data acquisition program, thus making it potentially suitable for customized applications.

  3. A repository based on a dynamically extensible data model supporting multidisciplinary research in neuroscience

    PubMed Central

    2012-01-01

    Background Robust, extensible and distributed databases integrating clinical, imaging and molecular data represent a substantial challenge for modern neuroscience. It is even more difficult to provide extensible software environments able to effectively target the rapidly changing data requirements and structures of research experiments. There is an increasing request from the neuroscience community for software tools addressing technical challenges about: (i) supporting researchers in the medical field to carry out data analysis using integrated bioinformatics services and tools; (ii) handling multimodal/multiscale data and metadata, enabling the injection of several different data types according to structured schemas; (iii) providing high extensibility, in order to address different requirements deriving from a large variety of applications simply through a user runtime configuration. Methods A dynamically extensible data structure supporting collaborative multidisciplinary research projects in neuroscience has been defined and implemented. We have considered extensibility issues from two different points of view. First, the improvement of data flexibility has been taken into account. This has been done through the development of a methodology for the dynamic creation and use of data types and related metadata, based on the definition of “meta” data model. This way, users are not constrainted to a set of predefined data and the model can be easily extensible and applicable to different contexts. Second, users have been enabled to easily customize and extend the experimental procedures in order to track each step of acquisition or analysis. This has been achieved through a process-event data structure, a multipurpose taxonomic schema composed by two generic main objects: events and processes. Then, a repository has been built based on such data model and structure, and deployed on distributed resources thanks to a Grid-based approach. Finally, data integration aspects have been addressed by providing the repository application with an efficient dynamic interface designed to enable the user to both easily query the data depending on defined datatypes and view all the data of every patient in an integrated and simple way. Results The results of our work have been twofold. First, a dynamically extensible data model has been implemented and tested based on a “meta” data-model enabling users to define their own data types independently from the application context. This data model has allowed users to dynamically include additional data types without the need of rebuilding the underlying database. Then a complex process-event data structure has been built, based on this data model, describing patient-centered diagnostic processes and merging information from data and metadata. Second, a repository implementing such a data structure has been deployed on a distributed Data Grid in order to provide scalability both in terms of data input and data storage and to exploit distributed data and computational approaches in order to share resources more efficiently. Moreover, data managing has been made possible through a friendly web interface. The driving principle of not being forced to preconfigured data types has been satisfied. It is up to users to dynamically configure the data model for the given experiment or data acquisition program, thus making it potentially suitable for customized applications. Conclusions Based on such repository, data managing has been made possible through a friendly web interface. The driving principle of not being forced to preconfigured data types has been satisfied. It is up to users to dynamically configure the data model for the given experiment or data acquisition program, thus making it potentially suitable for customized applications. PMID:23043673

  4. National Aeronautics and Space Administration Biological Specimen Repository

    NASA Technical Reports Server (NTRS)

    McMonigal, Kathleen A.; Pietrzyk, Robert a.; Johnson, Mary Anne

    2008-01-01

    The National Aeronautics and Space Administration Biological Specimen Repository (Repository) is a storage bank that is used to maintain biological specimens over extended periods of time and under well-controlled conditions. Samples from the International Space Station (ISS), including blood and urine, will be collected, processed and archived during the preflight, inflight and postflight phases of ISS missions. This investigation has been developed to archive biosamples for use as a resource for future space flight related research. The International Space Station (ISS) provides a platform to investigate the effects of microgravity on human physiology prior to lunar and exploration class missions. The storage of crewmember samples from many different ISS flights in a single repository will be a valuable resource with which researchers can study space flight related changes and investigate physiological markers. The development of the National Aeronautics and Space Administration Biological Specimen Repository will allow for the collection, processing, storage, maintenance, and ethical distribution of biosamples to meet goals of scientific and programmatic relevance to the space program. Archiving of the biosamples will provide future research opportunities including investigating patterns of physiological changes, analysis of components unknown at this time or analyses performed by new methodologies.

  5. SATORI: a system for ontology-guided visual exploration of biomedical data repositories.

    PubMed

    Lekschas, Fritz; Gehlenborg, Nils

    2018-04-01

    The ever-increasing number of biomedical datasets provides tremendous opportunities for re-use but current data repositories provide limited means of exploration apart from text-based search. Ontological metadata annotations provide context by semantically relating datasets. Visualizing this rich network of relationships can improve the explorability of large data repositories and help researchers find datasets of interest. We developed SATORI-an integrative search and visual exploration interface for the exploration of biomedical data repositories. The design is informed by a requirements analysis through a series of semi-structured interviews. We evaluated the implementation of SATORI in a field study on a real-world data collection. SATORI enables researchers to seamlessly search, browse and semantically query data repositories via two visualizations that are highly interconnected with a powerful search interface. SATORI is an open-source web application, which is freely available at http://satori.refinery-platform.org and integrated into the Refinery Platform. nils@hms.harvard.edu. Supplementary data are available at Bioinformatics online.

  6. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zavarin, Mavrik; Joseph, C.

    This progress report (Level 4 Milestone Number M4FT-16LL080303052) summarizes research conducted at Lawrence Livermore National Laboratory (LLNL) within the Crystalline Disposal R&D Activity Number FT-16LL080303051 and Crystalline International Collaborations Activity Number FT-16LL080303061. The focus of this research is the interaction of radionuclides with Engineered Barrier System (EBS) and host rock materials at various physico-chemical conditions relevant to subsurface repository environments. They include both chemical and physical processes such as solubility, sorption, and diffusion. The colloid-facilitated transport effort focused on preparation of a draft manuscript summarizing the state of knowledge and parameterization of colloid-facilitated transport mechanisms in support of reactive transportmore » and performance assessment models for generic crystalline repositories. This draft manuscript is being submitted as a level 3 milestone with LANL as the primary author. LLNL’s contribution to that effort is summarized only briefly in the present report. A manuscript summarizing long-term U(VI) diffusion experiments through bentonite backfill material was recently accepted for publication; the contents of that manuscript are summarized in the present report. The Np(IV) diffusion experiments were started mid-year and are ongoing. The completion of these experiments is planned for early FY17. Our progress in quantifying Np(IV) diffusion in bentonite backfill is summarized in the present report. Our involvement with the NEA TDB project was summarized in a recent Argillite Disposal activity report. It is not included in this report.« less

  7. 10 CFR 60.51 - License amendment for permanent closure.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... description of the program for post-permanent closure monitoring of the geologic repository. (2) A detailed... postclosure controlled area and geologic repository operations area by monuments that have been designed... tests, experiments, and any other analyses relating to backfill of excavated areas, shaft sealing, waste...

  8. 10 CFR 60.51 - License amendment for permanent closure.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... description of the program for post-permanent closure monitoring of the geologic repository. (2) A detailed... postclosure controlled area and geologic repository operations area by monuments that have been designed... tests, experiments, and any other analyses relating to backfill of excavated areas, shaft sealing, waste...

  9. 10 CFR 60.51 - License amendment for permanent closure.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... description of the program for post-permanent closure monitoring of the geologic repository. (2) A detailed... postclosure controlled area and geologic repository operations area by monuments that have been designed... tests, experiments, and any other analyses relating to backfill of excavated areas, shaft sealing, waste...

  10. 10 CFR 60.51 - License amendment for permanent closure.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... description of the program for post-permanent closure monitoring of the geologic repository. (2) A detailed... postclosure controlled area and geologic repository operations area by monuments that have been designed... tests, experiments, and any other analyses relating to backfill of excavated areas, shaft sealing, waste...

  11. 10 CFR 60.51 - License amendment for permanent closure.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... description of the program for post-permanent closure monitoring of the geologic repository. (2) A detailed... postclosure controlled area and geologic repository operations area by monuments that have been designed... tests, experiments, and any other analyses relating to backfill of excavated areas, shaft sealing, waste...

  12. Malaysian Education Index (MEI): An Online Indexing and Repository System

    ERIC Educational Resources Information Center

    Kabilan, Muhammad Kamarul; Ismail, Hairul Nizam; Yaakub, Rohizani; Yusof, Najeemah Mohd; Idros, Sharifah Noraidah Syed; Umar, Irfan Naufal; Arshad, Muhammad Rafie Mohd.; Idrus, Rosnah; Rahman, Habsah Abdul

    2010-01-01

    This "Project Sheet" describes an on-going project that is being carried out by a group of educational researchers, computer science researchers and librarians from Universiti Sains Malaysia, Penang. The Malaysian Education Index (MEI) has two main functions--(1) Online Indexing System, and (2) Online Repository System. In this brief…

  13. Personalized reminiscence therapy M-health application for patients living with dementia: Innovating using open source code repository.

    PubMed

    Zhang, Melvyn W B; Ho, Roger C M

    2017-01-01

    Dementia is known to be an illness which brings forth marked disability amongst the elderly individuals. At times, patients living with dementia do also experience non-cognitive symptoms, and these symptoms include that of hallucinations, delusional beliefs as well as emotional liability, sexualized behaviours and aggression. According to the National Institute of Clinical Excellence (NICE) guidelines, non-pharmacological techniques are typically the first-line option prior to the consideration of adjuvant pharmacological options. Reminiscence and music therapy are thus viable options. Lazar et al. [3] previously performed a systematic review with regards to the utilization of technology to delivery reminiscence based therapy to individuals who are living with dementia and has highlighted that technology does have benefits in the delivery of reminiscence therapy. However, to date, there has been a paucity of M-health innovations in this area. In addition, most of the current innovations are not personalized for each of the person living with Dementia. Prior research has highlighted the utility for open source repository in bioinformatics study. The authors hoped to explain how they managed to tap upon and make use of open source repository in the development of a personalized M-health reminiscence therapy innovation for patients living with dementia. The availability of open source code repository has changed the way healthcare professionals and developers develop smartphone applications today. Conventionally, a long iterative process is needed in the development of native application, mainly because of the need for native programming and coding, especially so if the application needs to have interactive features or features that could be personalized. Such repository enables the rapid and cost effective development of application. Moreover, developers are also able to further innovate, as less time is spend in the iterative process.

  14. Generation and validation of a universal perinatal database and biospecimen repository: PeriBank.

    PubMed

    Antony, K M; Hemarajata, P; Chen, J; Morris, J; Cook, C; Masalas, D; Gedminas, M; Brown, A; Versalovic, J; Aagaard, K

    2016-11-01

    There is a dearth of biospecimen repositories available to perinatal researchers. In order to address this need, here we describe the methodology used to establish such a resource. With the collaboration of MedSci.net, we generated an online perinatal database with 847 fields of clinical information. Simultaneously, we established a biospecimen repository of the same clinical participants. The demographic and clinical outcomes data are described for the first 10 000 participants enrolled. The demographic characteristics are consistent with the demographics of the delivery hospitals. Quality analysis of the biospecimens reveals variation in very few analytes. Furthermore, since the creation of PeriBank, we have demonstrated validity of the database and tissue integrity of the biospecimen repository. Here we establish that the creation of a universal perinatal database and biospecimen collection is not only possible, but allows for the performance of state-of-the-science translational perinatal research and is a potentially valuable resource to academic perinatal researchers.

  15. DNASU plasmid and PSI:Biology-Materials repositories: resources to accelerate biological research

    PubMed Central

    Seiler, Catherine Y.; Park, Jin G.; Sharma, Amit; Hunter, Preston; Surapaneni, Padmini; Sedillo, Casey; Field, James; Algar, Rhys; Price, Andrea; Steel, Jason; Throop, Andrea; Fiacco, Michael; LaBaer, Joshua

    2014-01-01

    The mission of the DNASU Plasmid Repository is to accelerate research by providing high-quality, annotated plasmid samples and online plasmid resources to the research community through the curated DNASU database, website and repository (http://dnasu.asu.edu or http://dnasu.org). The collection includes plasmids from grant-funded, high-throughput cloning projects performed in our laboratory, plasmids from external researchers, and large collections from consortia such as the ORFeome Collaboration and the NIGMS-funded Protein Structure Initiative: Biology (PSI:Biology). Through DNASU, researchers can search for and access detailed information about each plasmid such as the full length gene insert sequence, vector information, associated publications, and links to external resources that provide additional protein annotations and experimental protocols. Plasmids can be requested directly through the DNASU website. DNASU and the PSI:Biology-Materials Repositories were previously described in the 2010 NAR Database Issue (Cormier, C.Y., Mohr, S.E., Zuo, D., Hu, Y., Rolfs, A., Kramer, J., Taycher, E., Kelley, F., Fiacco, M., Turnbull, G. et al. (2010) Protein Structure Initiative Material Repository: an open shared public resource of structural genomics plasmids for the biological community. Nucleic Acids Res., 38, D743–D749.). In this update we will describe the plasmid collection and highlight the new features in the website redesign, including new browse/search options, plasmid annotations and a dynamic vector mapping feature that was developed in collaboration with LabGenius. Overall, these plasmid resources continue to enable research with the goal of elucidating the role of proteins in both normal biological processes and disease. PMID:24225319

  16. Institutional Repositories: The Experience of Master's and Baccalaureate Institutions

    ERIC Educational Resources Information Center

    Markey, Karen; St. Jean, Beth; Soo, Young Rieh; Yakel, Elizabeth; Kim, Jihyun

    2008-01-01

    In 2006, MIRACLE Project investigators censused library directors at all U.S. academic institutions about their activities planning, pilot testing, and implementing the institutional repositories on their campuses. Out of 446 respondents, 289 (64.8 percent) were from master's and baccalaureate institutions (M&BIs) where few operational…

  17. Indian Institutional Repositories: A Study of User's Perspective

    ERIC Educational Resources Information Center

    Sawant, Sarika

    2012-01-01

    Purpose: The present study aims to investigate the experience, contribution and opinions of users of respective institutional repositories (IRs) developed in India. Design/methodology/approach: The survey method was used. The data collection tool was a web questionnaire, which was created with the help of software provided by surveymonkey.com…

  18. Collaborative Recommendation of E-Learning Resources: An Experimental Investigation

    ERIC Educational Resources Information Center

    Manouselis, N.; Vuorikari, R.; Van Assche, F.

    2010-01-01

    Repositories with educational resources can support the formation of online learning communities by providing a platform for collaboration. Users (e.g. teachers, tutors and learners) access repositories, search for interesting resources to access and use, and in many cases, also exchange experiences and opinions. A particular class of online…

  19. Accessing and Integrating Data and Knowledge for Biomedical Research

    PubMed Central

    Burgun, A.; Bodenreider, O.

    2008-01-01

    Summary Objectives To review the issues that have arisen with the advent of translational research in terms of integration of data and knowledge, and survey current efforts to address these issues. Methods Using examples form the biomedical literature, we identified new trends in biomedical research and their impact on bioinformatics. We analyzed the requirements for effective knowledge repositories and studied issues in the integration of biomedical knowledge. Results New diagnostic and therapeutic approaches based on gene expression patterns have brought about new issues in the statistical analysis of data, and new workflows are needed are needed to support translational research. Interoperable data repositories based on standard annotations, infrastructures and services are needed to support the pooling and meta-analysis of data, as well as their comparison to earlier experiments. High-quality, integrated ontologies and knowledge bases serve as a source of prior knowledge used in combination with traditional data mining techniques and contribute to the development of more effective data analysis strategies. Conclusion As biomedical research evolves from traditional clinical and biological investigations towards omics sciences and translational research, specific needs have emerged, including integrating data collected in research studies with patient clinical data, linking omics knowledge with medical knowledge, modeling the molecular basis of diseases, and developing tools that support in-depth analysis of research data. As such, translational research illustrates the need to bridge the gap between bioinformatics and medical informatics, and opens new avenues for biomedical informatics research. PMID:18660883

  20. The Pig PeptideAtlas: A resource for systems biology in animal production and biomedicine.

    PubMed

    Hesselager, Marianne O; Codrea, Marius C; Sun, Zhi; Deutsch, Eric W; Bennike, Tue B; Stensballe, Allan; Bundgaard, Louise; Moritz, Robert L; Bendixen, Emøke

    2016-02-01

    Biological research of Sus scrofa, the domestic pig, is of immediate relevance for food production sciences, and for developing pig as a model organism for human biomedical research. Publicly available data repositories play a fundamental role for all biological sciences, and protein data repositories are in particular essential for the successful development of new proteomic methods. Cumulative proteome data repositories, including the PeptideAtlas, provide the means for targeted proteomics, system-wide observations, and cross-species observational studies, but pigs have so far been underrepresented in existing repositories. We here present a significantly improved build of the Pig PeptideAtlas, which includes pig proteome data from 25 tissues and three body fluid types mapped to 7139 canonical proteins. The content of the Pig PeptideAtlas reflects actively ongoing research within the veterinary proteomics domain, and this article demonstrates how the expression of isoform-unique peptides can be observed across distinct tissues and body fluids. The Pig PeptideAtlas is a unique resource for use in animal proteome research, particularly biomarker discovery and for preliminary design of SRM assays, which are equally important for progress in research that supports farm animal production and veterinary health, as for developing pig models with relevance to human health research. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  1. The Pig PeptideAtlas: a resource for systems biology in animal production and biomedicine

    PubMed Central

    Hesselager, Marianne O.; Codrea, Marius C.; Sun, Zhi; Deutsch, Eric W.; Bennike, Tue B.; Stensballe, Allan; Bundgaard, Louise; Moritz, Robert L.; Bendixen, Emøke

    2016-01-01

    Biological research of Sus scrofa, the domestic pig, is of immediate relevance for food production sciences, and for developing pig as a model organism for human biomedical research. Publicly available data repositories play a fundamental role for all biological sciences, and protein data repositories are in particular essential for the successful development of new proteomic methods. Cumulative proteome data repositories, including the PeptideAtlas, provide the means for targeted proteomics, system wide observations, and cross species observational studies, but pigs have so far been underrepresented in existing repositories. We here present a significantly improved build of the Pig PeptideAtlas, which includes pig proteome data from 25 tissues and three body fluid types mapped to 7139 canonical proteins. The content of the Pig PeptideAtlas reflects actively ongoing research within the veterinary proteomics domain, and this manuscript demonstrates how the expression of isoform-unique peptides can be observed across distinct tissues and body fluids. The Pig PeptideAtlas is a unique resource for use in animal proteome research, particularly biomarker discovery and for preliminary design of SRM assays, which are equally important for progress in research that supports farm animal production and veterinary health, as for developing pig models with relevance to human health research. PMID:26699206

  2. Promoting Academic Physicists, Their Students, and Their Research through Library Partnerships

    NASA Astrophysics Data System (ADS)

    Rozum, B.; Wesolek, A.

    2012-12-01

    At many institutions, attracting and mentoring quality students is of key importance. Through their developing careers, typically under the tutelage of one primary faculty member, students build portfolios, prepare for graduate school, and apply to post-doc programs or faculty positions. Often though, the corpus of that primary faculty member's work is not available in a single location. This is a disadvantage both for current students, who wish to highlight the importance of their work within the context of a research group and for the department, which can miss opportunities to attract high-quality future students. Utah State University Libraries hosts a thriving institutional repository, DigitalCommons@USU, which provides open access to scholarly works, research, reports, publications, and journals produced by Utah State University faculty, staff, and students. The Library and the Physics Department developed a partnership to transcend traditional library repository architecture and emphasize faculty research groups within the department. Previously, only student theses and dissertations were collected, and they were not associated with the department in any way. Now student presentations, papers, and posters appear with other faculty works all in the same research work space. This poster session highlights the features of the University's repository and describes what is required to establish a similar structure at other academic institutions. We anticipate several long-term benefits of this new structure. Students are pleased with the increased visibility of their research and with having an online presence through their "Selected Works" personal author site. Faculty are pleased with the opportunity to highlight their research and the potential to attract new students to their research groups. This new repository model also allows the library to amplify the existing scientific outreach initiatives of the physics department. One example of this is a recent exhibit created in the Library showcasing a student research group's 30-year history of sending payloads into space. The exhibit was a direct result of archiving the work of student researchers in the institutional repository. From the perspective of the Library, the benefits are also impressive. The Library is able to build its institutional repository, develop strong relations with faculty in the Physics Department, and have access to unpublished reports that otherwise might be lost. Establishing research groups' presence in DigitalCommons@USU provided an opportunity to meet with the Physics graduate students to discuss setting up online web portfolios, archiving their publications, and understanding publisher contracts. Developing partnerships between academic units and libraries is one more method to reach out to potential students, promote research, and showcase the talents of faculty and students. Using the Library's institutional repository to do this is beneficial for everyone.

  3. Leading across Boundaries: Collaborative Leadership and the Institutional Repository in Research Universities and Liberal Arts Colleges

    ERIC Educational Resources Information Center

    Seaman, David M.

    2017-01-01

    Libraries often engage in services that require collaboration across stakeholder boundaries to be successful. Institutional repositories (IRs) are a good example of such a service. IRs are an infrastructure to preserve intellectual assets within a university or college, and to provide an open access showcase for that institution's research,…

  4. Collaboration, Coherence and Capacity-Building: The Role of DSpace in Supporting and Understanding the TLRP

    ERIC Educational Resources Information Center

    Procter, Richard

    2007-01-01

    This paper describes how the Teaching and Learning Research Programme (TLRP) has implemented and applied DSpace as a digital repository for project and programme outputs, including published articles, conference papers, research reports, briefings and press releases. The DSpace repository has become a major element in the user engagement strategy…

  5. Use of Digital Repositories by Chemistry Researchers: Results of a Survey

    ERIC Educational Resources Information Center

    Polydoratou, Panayiota

    2007-01-01

    Purpose: This paper aims to present findings from a survey that aimed to identify the issues around the use and linkage of source and output repositories and the chemistry researchers' expectations about their use. Design/methodology/approach: This survey was performed by means of an online questionnaire and structured interviews with academic and…

  6. Open Access to Physics and Astronomy Theses: A Case Study of the Raman Research Institute Digital Repository

    NASA Astrophysics Data System (ADS)

    Nagaraj, M. N.; Manjunath, M.; Savanur, K. P.; Sheshadri, G.

    2010-10-01

    With the introduction of information technology (IT) and its applications, libraries have started looking for ways to promote their institutes' research output. At the Raman Research Institute (RRI), we have showcased research output such as research papers, newspaper clippings, annual reports, technical reports, and the entire collection of C.V. Raman through the RRI digital repository, using DSpace. Recently, we have added doctoral dissertations to the repository and have made them accessible with the author's permission. In this paper, we describe the challenges and problems encountered in this project. The various stages including policy decisions, the scanning process, getting permissions, metadata standards and other related issues are described. We conclude by making a plea to other institutions also to make their theses available open-access so that this valuable information resource is accessible to all.

  7. The Materials Commons: A Collaboration Platform and Information Repository for the Global Materials Community

    NASA Astrophysics Data System (ADS)

    Puchala, Brian; Tarcea, Glenn; Marquis, Emmanuelle. A.; Hedstrom, Margaret; Jagadish, H. V.; Allison, John E.

    2016-08-01

    Accelerating the pace of materials discovery and development requires new approaches and means of collaborating and sharing information. To address this need, we are developing the Materials Commons, a collaboration platform and information repository for use by the structural materials community. The Materials Commons has been designed to be a continuous, seamless part of the scientific workflow process. Researchers upload the results of experiments and computations as they are performed, automatically where possible, along with the provenance information describing the experimental and computational processes. The Materials Commons website provides an easy-to-use interface for uploading and downloading data and data provenance, as well as for searching and sharing data. This paper provides an overview of the Materials Commons. Concepts are also outlined for integrating the Materials Commons with the broader Materials Information Infrastructure that is evolving to support the Materials Genome Initiative.

  8. The NASA Ames Life Sciences Data Archive: Biobanking for the Final Frontier

    NASA Technical Reports Server (NTRS)

    Rask, Jon; Chakravarty, Kaushik; French, Alison J.; Choi, Sungshin; Stewart, Helen J.

    2017-01-01

    The NASA Ames Institutional Scientific Collection involves the Ames Life Sciences Data Archive (ALSDA) and a biospecimen repository, which are responsible for archiving information and non-human biospecimens collected from spaceflight and matching ground control experiments. The ALSDA also manages a biospecimen sharing program, performs curation and long-term storage operations, and facilitates distribution of biospecimens for research purposes via a public website (https:lsda.jsc.nasa.gov). As part of our best practices, a tissue viability testing plan has been developed for the repository, which will assess the quality of samples subjected to long-term storage. We expect that the test results will confirm usability of the samples, enable broader science community interest, and verify operational efficiency of the archives. This work will also support NASA open science initiatives and guides development of NASA directives and policy for curation of biological collections.

  9. Tracking Research Data Footprints via Integration with Research Graph

    NASA Astrophysics Data System (ADS)

    Evans, B. J. K.; Wang, J.; Aryani, A.; Conlon, M.; Wyborn, L. A.; Choudhury, S. A.

    2017-12-01

    The researcher of today is likely to be part of a team that will use subsets of data from at least one, if not more external repositories, and that same data could be used by multiple researchers for many different purposes. At best, the repositories that host this data will know who is accessing their data, but rarely what they are using it for, resulting in funders of data collecting programs and data repositories that store the data unlikely to know: 1) which research funding contributed to the collection and preservation of a dataset, and 2) which data contributed to high impact research and publications. In days of funding shortages there is a growing need to be able to trace the footprint a data set from the originator that collected the data to the repository that stores the data and ultimately to any derived publications. The Research Data Alliance's Data Description Registry Interoperability Working Group (DDRIWG) has addressed this problem through the development of a distributed graph, called Research Graph that can map each piece of the research interaction puzzle by building aggregated graphs. It can connect datasets on the basis of co-authorship or other collaboration models such as joint funding and grants and can connect research datasets, publications, grants and researcher profiles across research repositories and infrastructures such as DataCite and ORCID. National Computational Infrastructure (NCI) in Australia is one of the early adopters of Research Graph. The graphic view and quantitative analysis helps NCI track the usage of their National reference data collections thus quantifying the role that these NCI-hosted data assets play within the funding-researcher-data-publication-cycle. The graph can unlock the complex interactions of the research projects by tracking the contribution of datasets, the various funding bodies and the downstream data users. RMap Project is a similar initiative which aims to solve complex relationships among scholarly publications and their underlying data, including IEEE publications. It is hoped to combine RMap and Research Graph in the near futures and also to add physical samples to Research Graph.

  10. Accelerating research through reagent repositories: the genome editing example.

    PubMed

    Joung, J Keith; Voytas, Daniel F; Kamens, Joanne

    2015-11-20

    Keith Joung, Dan Voytas and Joanne Kamens share insights into how the genome editing field was advanced by early access to biological resources and the role in this process that plasmid repositories play.

  11. Migration of the Gaudi and LHCb software repositories from CVS to Subversion

    NASA Astrophysics Data System (ADS)

    Clemencic, M.; Degaudenzi, H.; LHCb Collaboration

    2011-12-01

    A common code repository is of primary importance in a distributed development environment such as large HEP experiments. CVS (Concurrent Versions System) has been used in the past years at CERN for the hosting of shared software repositories, among which were the repositories for the Gaudi Framework and the LHCb software projects. Many developers around the world produced alternative systems to share code and revisions among several developers, mainly to overcome the limitations in CVS, and CERN has recently started a new service for code hosting based on the version control system Subversion. The differences between CVS and Subversion and the way the code was organized in Gaudi and LHCb CVS repositories required careful study and planning of the migration. Special care was used to define the organization of the new Subversion repository. To avoid as much as possible disruption in the development cycle, the migration has been gradual with the help of tools developed explicitly to hide the differences between the two systems. The principles guiding the migration steps, the organization of the Subversion repository and the tools developed will be presented, as well as the problems encountered both from the librarian and the user points of view.

  12. Adapting a Clinical Data Repository to ICD-10-CM through the use of a Terminology Repository

    PubMed Central

    Cimino, James J.; Remennick, Lyubov

    2014-01-01

    Clinical data repositories frequently contain patient diagnoses coded with the International Classification of Diseases, Ninth Revision (ICD-9-CM). These repositories now need to accommodate data coded with the Tenth Revision (ICD-10-CM). Database users wish to retrieve relevant data regardless of the system by which they are coded. We demonstrate how a terminology repository (the Research Entities Dictionary or RED) serves as an ontology relating terms of both ICD versions to each other to support seamless version-independent retrieval from the Biomedical Translational Research Information System (BTRIS) at the National Institutes of Health. We make use of the Center for Medicare and Medicaid Services’ General Equivalence Mappings (GEMs) to reduce the modeling effort required to determine whether ICD-10-CM terms should be added to the RED as new concepts or as synonyms of existing concepts. A divide-and-conquer approach is used to develop integration heuristics that offer a satisfactory interim solution and facilitate additional refinement of the integration as time and resources allow. PMID:25954344

  13. Investigation of Coupled Processes and Impact of High Temperature Limits in Argillite Rock: FY17 Progress. Predecisional Draft

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zheng, Liange; Rutqvist, Jonny; Xu, Hao

    The focus of research within the Spent Fuel and Waste Science and Technology (SFWST) (formerly called Used Fuel Disposal) Campaign is on repository-induced interactions that may affect the key safety characteristics of EBS bentonite and an argillaceous rock. These include thermal-hydrologicalmechanical- chemical (THMC) process interactions that occur as a result of repository construction and waste emplacement. Some of the key questions addressed in this report include the development of fracturing in the excavation damaged zone (EDZ) and THMC effects on the near-field argillaceous rock and buffer materials and petrophysical characteristics, particularly the impacts of temperature rise caused by waste heat.more » This report documents the following research activities. Section 2 presents THM model developments and validation, including modeling of underground heater experiments at Mont Terri and Bure underground research laboratories (URLs). The heater experiments modeled are the Mont Terri FE (Full-scale Emplacement) Experiment, conducted as part of the Mont Terri Project, and the TED in heater test conducted in Callovo-Oxfordian claystone (COx) at the Meuse/Haute-Marne (MHM) underground research laboratory in France. The modeling of the TED heater test is one of the Tasks of the DEvelopment of COupled Models and their VAlidation against EXperiments (DECOVALEX)-2019 project. Section 3 presents the development and application of thermal-hydrological-mechanical-chemical (THMC) modeling to evaluate EBS bentonite and argillite rock responses under different temperatures (100 °C and 200 °C). Model results are presented to help to understand the impact of high temperatures on the properties and behavior of bentonite and argillite rock. Eventually the process model will support a robust GDSA model for repository performance assessments. Section 4 presents coupled THMC modeling for an in situ test conducted at Grimsel underground laboratory in Switzerland in the Full-Scale Engineered Barrier Experiment Dismantling Project (FEBEX-DP). The data collected in the test after almost two decades of heating and two dismantling events provide a unique opportunity of validating coupled THMC models and enhancing our understanding of coupled THMC process in EBS bentonite. Section 5 presents a planned large in-situ test, “HotBENT,” at Grimsel Test Site, Switzerland. In this test, bentonite backfilled EBS in granite will be heated up to 200 °C, where the most relevant features of future emplacement conditions can be adequately reproduced. Lawrence Berkeley National Laboratory (LBNL) has very actively participated in the project since the very beginning and have conducted scoping calculations in FY17 to facilitate the final design of the experiment. Section 6 presents present LBNL’s activities for modeling gas migration in clay related to Task A of the international DECOVALEX-2019 project. This is an international collaborative activity in which DOE and LBNL gain access to unique laboratory and field data of gas migration that are studied with numerical modeling to better understand the processes, to improve numerical models that could eventually be applied in the performance assessment for nuclear waste disposal in clay host rocks and bentonite backfill. Section 7 summarizes the main research accomplishments for FY17 and proposes future work activities.« less

  14. [The subject repositories of strategy of the Open Access initiative].

    PubMed

    Soares Guimarães, M C; da Silva, C H; Horsth Noronha, I

    2012-11-01

    The subject repositories are defined as a set of digital objects resulting from the research related to a specific disciplinary field and occupy a still restricted space in the discussion agenda of the Free Access Movement when compared to amplitude reached in the discussion of Institutional Repositories. Although the Subject Repository comes to prominence in the field, especially for the success of initiatives such as the arXiv, PubMed and E-prints, the literature on the subject is recognized as very limited. Despite its roots in the Library and Information Science, and focus on the management of disciplinary collections (subject area literature), there is little information available about the development and management of subject repositories. The following text seeks to make a brief summary on the topic as a way to present the potential to develop subject repositories in order to strengthen the initiative of open access.

  15. Databases and Web Tools for Cancer Genomics Study

    PubMed Central

    Yang, Yadong; Dong, Xunong; Xie, Bingbing; Ding, Nan; Chen, Juan; Li, Yongjun; Zhang, Qian; Qu, Hongzhu; Fang, Xiangdong

    2015-01-01

    Publicly-accessible resources have promoted the advance of scientific discovery. The era of genomics and big data has brought the need for collaboration and data sharing in order to make effective use of this new knowledge. Here, we describe the web resources for cancer genomics research and rate them on the basis of the diversity of cancer types, sample size, omics data comprehensiveness, and user experience. The resources reviewed include data repository and analysis tools; and we hope such introduction will promote the awareness and facilitate the usage of these resources in the cancer research community. PMID:25707591

  16. Geoscience Digital Data Resource and Repository Service

    NASA Astrophysics Data System (ADS)

    Mayernik, M. S.; Schuster, D.; Hou, C. Y.

    2017-12-01

    The open availability and wide accessibility of digital data sets is becoming the norm for geoscience research. The National Science Foundation (NSF) instituted a data management planning requirement in 2011, and many scientific publishers, including the American Geophysical Union and the American Meteorological Society, have recently implemented data archiving and citation policies. Many disciplinary data facilities exist around the community to provide a high level of technical support and expertise for archiving data of particular kinds, or for particular projects. However, a significant number of geoscience research projects do not have the same level of data facility support due to a combination of several factors, including the research project's size, funding limitations, or topic scope that does not have a clear facility match. These projects typically manage data on an ad hoc basis without limited long-term management and preservation procedures. The NSF is supporting a workshop to be held in Summer of 2018 to develop requirements and expectations for a Geoscience Digital Data Resource and Repository Service (GeoDaRRS). The vision for the prospective GeoDaRRS is to complement existing NSF-funded data facilities by providing: 1) data management planning support resources for the general community, and 2) repository services for researchers who have data that do not fit in any existing repository. Functionally, the GeoDaRRS would support NSF-funded researchers in meeting data archiving requirements set by the NSF and publishers for geosciences, thereby ensuring the availability of digital data for use and reuse in scientific research going forward. This presentation will engage the AGU community in discussion about the needs for a new digital data repository service, specifically to inform the forthcoming GeoDaRRS workshop.

  17. Academic Research Library as Broker in Addressing Interoperability Challenges for the Geosciences

    NASA Astrophysics Data System (ADS)

    Smith, P., II

    2015-12-01

    Data capture is an important process in the research lifecycle. Complete descriptive and representative information of the data or database is necessary during data collection whether in the field or in the research lab. The National Science Foundation's (NSF) Public Access Plan (2015) mandates the need for federally funded projects to make their research data more openly available. Developing, implementing, and integrating metadata workflows into to the research process of the data lifecycle facilitates improved data access while also addressing interoperability challenges for the geosciences such as data description and representation. Lack of metadata or data curation can contribute to (1) semantic, (2) ontology, and (3) data integration issues within and across disciplinary domains and projects. Some researchers of EarthCube funded projects have identified these issues as gaps. These gaps can contribute to interoperability data access, discovery, and integration issues between domain-specific and general data repositories. Academic Research Libraries have expertise in providing long-term discovery and access through the use of metadata standards and provision of access to research data, datasets, and publications via institutional repositories. Metadata crosswalks, open archival information systems (OAIS), trusted-repositories, data seal of approval, persistent URL, linking data, objects, resources, and publications in institutional repositories and digital content management systems are common components in the library discipline. These components contribute to a library perspective on data access and discovery that can benefit the geosciences. The USGS Community for Data Integration (CDI) has developed the Science Support Framework (SSF) for data management and integration within its community of practice for contribution to improved understanding of the Earth's physical and biological systems. The USGS CDI SSF can be used as a reference model to map to EarthCube Funded projects with academic research libraries facilitating the data and information assets components of the USGS CDI SSF via institutional repositories and/or digital content management. This session will explore the USGS CDI SSF for cross-discipline collaboration considerations from a library perspective.

  18. Benefits of International Collaboration on the International Space Station

    NASA Technical Reports Server (NTRS)

    Robinson, Julie A.; Hasbrook, Pete; Tate Brown, Judy; Thumm, Tracy; Cohen, Luchino; Marcil, Isabelle; De Parolis, Lina; Hatton, Jason; Umezawa, Kazuo; Shirakawa, Masaki; hide

    2017-01-01

    The International Space Station is a valuable platform for research in space, but the benefits are limited if research is only conducted by individual countries. Through the e orts of the ISS Program Science Forum, international science working groups, and interagency cooperation, international collaboration on the ISS has expanded as ISS utilization has matured. Members of science teams benefit from working with counterparts in other countries. Scientists and institutions bring years of experience and specialized expertise to collaborative investigations, leading to new perspectives and approaches to scientific challenges. Combining new ideas and historical results brings synergy and improved peer-reviewed scientific methods and results. World-class research facilities can be expensive and logistically complicated, jeopardizing their full utilization. Experiments that would be prohibitively expensive for a single country can be achieved through contributions of resources from two or more countries, such as crew time, up- and downmass, and experiment hardware. Cooperation also avoids duplication of experiments and hardware among agencies. Biomedical experiments can be completed earlier if astronauts or cosmonauts from multiple agencies participate. Countries responding to natural disasters benefit from ISS imagery assets, even if the country has no space agency of its own. Students around the world participate in ISS educational opportunities, and work with students in other countries, through open curriculum packages and through international competitions. Even experiments conducted by a single country can benefit scientists around the world, through specimen sharing programs and publicly accessible \\open data" repositories. For ISS data, these repositories include GeneLab, the Physical Science Informatics System, and different Earth science data systems. Scientists can conduct new research using ISS data without having to launch and execute their own experiments. Multilateral collections of research results publications, maintained by the ISS international partnership and accessible via nasa.gov, make ISS results available worldwide, and encourage new users, ideas and research. The paper explores effectiveness of international collaboration in the course of the ISS Program execution. The collaboration history, its evolution and maturation, change of focus during its different phases, and growth of its effectiveness (in accordance with the especially established criteria) are also considered in the paper in the light of benefits for the entire ISS community. With the International Space Station extended through at least 2024, more crew time becoming available and new facilities arriving on board the ISS, these benefits of international scientific collaboration on the ISS can only increase.

  19. Building and Using Digital Repository Certifications across Science

    NASA Astrophysics Data System (ADS)

    McIntosh, L.

    2017-12-01

    When scientific recommendations are made based upon research, the quality and integrity of the data should be rigorous enough to verify claims and in a trusted location. Key to ensuring the transparency and verifiability of research, reproducibility hinges not only on the availability of the documentation, analyses, and data, but the ongoing accessibility and viability of the files and documents, enhanced through a process of curation. The Research Data Alliance (RDA) is an international, community-driven, action-oriented, virtual organization committed to enabling the open sharing of data by building social and technical bridges. Within the RDA, multiple groups are working on consensus-building around the certification of digital repositories across scientific domains. For this section of the panel, we will discuss the work to date on repository certification from this RDA perspective.

  20. DataUp 2.0: Improving On a Tool For Helping Researchers Archive, Manage, and Share Their Tabular Data

    NASA Astrophysics Data System (ADS)

    Strasser, C.; Borda, S.; Cruse, P.; Kunze, J.

    2013-12-01

    There are many barriers to data management and sharing among earth and environmental scientists; among the most significant are a lack of knowledge about best practices for data management, metadata standards, or appropriate data repositories for archiving and sharing data. Last year we developed an open source web application, DataUp, to help researchers overcome these barriers. DataUp helps scientists to (1) determine whether their file is CSV compatible, (2) generate metadata in a standard format, (3) retrieve an identifier to facilitate data citation, and (4) deposit their data into a repository. With funding from the NSF via a supplemental grant to the DataONE project, we are working to improve upon DataUp. Our main goal for DataUp 2.0 is to ensure organizations and repositories are able to adopt and adapt DataUp to meet their unique needs, including connecting to analytical tools, adding new metadata schema, and expanding the list of connected data repositories. DataUp is a collaborative project between the California Digital Library, DataONE, the San Diego Supercomputing Center, and Microsoft Research Connections.

  1. Cancer Epidemiology Data Repository (CEDR)

    Cancer.gov

    In an effort to broaden access and facilitate efficient data sharing, the Epidemiology and Genomics Research Program (EGRP) has created the Cancer Epidemiology Data Repository (CEDR), a centralized, controlled-access database, where Investigators can deposit individual-level de-identified observational cancer datasets.

  2. Repository contributions to Rubus research

    USDA-ARS?s Scientific Manuscript database

    The USDA National Plant Germplasm System is a nation-wide source for global genetic resources. The National Clonal Germplasm Repository (NCGR) in Corvallis, OR, maintains crops and crop wild relatives for the Willamette Valley including pear, raspberry and blackberry, strawberry, blueberry, gooseber...

  3. Create a translational medicine knowledge repository--research downsizing, mergers and increased outsourcing have reduced the depth of in-house translational medicine expertise and institutional memory at many pharmaceutical and biotech companies: how will they avoid relearning old lessons?

    PubMed

    Littman, Bruce H; Marincola, Francesco M

    2011-05-10

    Pharmaceutical industry consolidation and overall research downsizing threatens the ability of companies to benefit from their previous investments in translational research as key leaders with the most knowledge of the successful use of biomarkers and translational pharmacology models are laid off or accept their severance packages. Two recently published books may help to preserve this type of knowledge but much of this type of information is not in the public domain. Here we propose the creation of a translational medicine knowledge repository where companies can submit their translational research data and access similar data from other companies in a precompetitive environment. This searchable repository would become an invaluable resource for translational scientists and drug developers that could speed and reduce the cost of new drug development.

  4. Proceedings of the 7th US/German Workshop on Salt Repository Research, Design, and Operation.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hansen, Francis D.; Steininger, Walter; Bollingerfehr, Willhelm

    The 7th US/German Workshop on Salt Repository Research, Design, and Operation was held in Washington, DC on September 7-9, 2016. Over fifty participants representing governmental agencies, internationally recognized salt research groups, universities, and private companies helped advance the technical basis for salt disposal of radioactive waste. Representatives from several United States federal agencies were able to attend, including the Department of Energy´s Office of Environmental Management and Office of Nuclear Energy, the Environmental Protection Agency, the Nuclear Regulatory Commission, and the Nuclear Waste Technical Review Board. A similar representation from the German ministries showcased the covenant established in a Memorandummore » of Understanding executed between the United States and Germany in 2011. The US/German workshops´ results and activities also contribute significantly to the Nuclear Energy Agency Salt Club repository research agenda.« less

  5. Create a translational medicine knowledge repository - Research downsizing, mergers and increased outsourcing have reduced the depth of in-house translational medicine expertise and institutional memory at many pharmaceutical and biotech companies: how will they avoid relearning old lessons?

    PubMed Central

    2011-01-01

    Pharmaceutical industry consolidation and overall research downsizing threatens the ability of companies to benefit from their previous investments in translational research as key leaders with the most knowledge of the successful use of biomarkers and translational pharmacology models are laid off or accept their severance packages. Two recently published books may help to preserve this type of knowledge but much of this type of information is not in the public domain. Here we propose the creation of a translational medicine knowledge repository where companies can submit their translational research data and access similar data from other companies in a precompetitive environment. This searchable repository would become an invaluable resource for translational scientists and drug developers that could speed and reduce the cost of new drug development. PMID:21569250

  6. Warehousing re-annotated cancer genes for biomarker meta-analysis.

    PubMed

    Orsini, M; Travaglione, A; Capobianco, E

    2013-07-01

    Translational research in cancer genomics assigns a fundamental role to bioinformatics in support of candidate gene prioritization with regard to both biomarker discovery and target identification for drug development. Efforts in both such directions rely on the existence and constant update of large repositories of gene expression data and omics records obtained from a variety of experiments. Users who interactively interrogate such repositories may have problems in retrieving sample fields that present limited associated information, due for instance to incomplete entries or sometimes unusable files. Cancer-specific data sources present similar problems. Given that source integration usually improves data quality, one of the objectives is keeping the computational complexity sufficiently low to allow an optimal assimilation and mining of all the information. In particular, the scope of integrating intraomics data can be to improve the exploration of gene co-expression landscapes, while the scope of integrating interomics sources can be that of establishing genotype-phenotype associations. Both integrations are relevant to cancer biomarker meta-analysis, as the proposed study demonstrates. Our approach is based on re-annotating cancer-specific data available at the EBI's ArrayExpress repository and building a data warehouse aimed to biomarker discovery and validation studies. Cancer genes are organized by tissue with biomedical and clinical evidences combined to increase reproducibility and consistency of results. For better comparative evaluation, multiple queries have been designed to efficiently address all types of experiments and platforms, and allow for retrieval of sample-related information, such as cell line, disease state and clinical aspects. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.

  7. New directions in medical e-curricula and the use of digital repositories.

    PubMed

    Fleiszer, David M; Posel, Nancy H; Steacy, Sean P

    2004-03-01

    Medical educators involved in the growth of multimedia-enhanced e-curricula are increasingly aware of the need for digital repositories to catalogue, store and ensure access to learning objects that are integrated within their online material. The experience at the Faculty of Medicine at McGill University during initial development of a mainstream electronic curriculum reflects this growing recognition that repositories can facilitate the development of a more comprehensive as well as effective electronic curricula. Also, digital repositories can help to ensure efficient utilization of resources through the use, re-use, and reprocessing of multimedia learning, addressing the potential for collaboration among repositories and increasing available material exponentially. The authors review different approaches to the development of a digital repository application, as well as global and specific issues that should be examined in the initial requirements definition and development phase, to ensure current initiatives meet long-term requirements. Often, decisions regarding creation of e-curricula and associated digital repositories are left to interested faculty and their individual development teams. However, the development of an e-curricula and digital repository is not predominantly a technical exercise, but rather one that affects global pedagogical strategies and curricular content and involves a commitment of large-scale resources. Outcomes of these decisions can have long-term consequences and as such, should involve faculty at the highest levels including the dean.

  8. Optimizing Decision Preparedness by Adapting Scenario Complexity and Automating Scenario Generation

    NASA Technical Reports Server (NTRS)

    Dunne, Rob; Schatz, Sae; Flore, Stephen M.; Nicholson, Denise

    2011-01-01

    Klein's recognition-primed decision (RPD) framework proposes that experts make decisions by recognizing similarities between current decision situations and previous decision experiences. Unfortunately, military personnel arQ often presented with situations that they have not experienced before. Scenario-based training (S8T) can help mitigate this gap. However, SBT remains a challenging and inefficient training approach. To address these limitations, the authors present an innovative formulation of scenario complexity that contributes to the larger research goal of developing an automated scenario generation system. This system will enable trainees to effectively advance through a variety of increasingly complex decision situations and experiences. By adapting scenario complexities and automating generation, trainees will be provided with a greater variety of appropriately calibrated training events, thus broadening their repositories of experience. Preliminary results from empirical testing (N=24) of the proof-of-concept formula are presented, and future avenues of scenario complexity research are also discussed.

  9. The Luminosity Measurement for the DZERO Experiment at Fermilab

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Snow, Gregory R.

    Primary project objective: The addition of University of Nebraska-Lincoln (UNL) human resources supported by this grant helped ensure that Fermilab’s DZERO experiment had a reliable luminosity measurement through the end of Run II data taking and an easily-accessible repository of luminosity information for all collaborators performing physics analyses through the publication of its final physics results. Secondary project objective: The collaboration between the UNL Instrument Shop and Fermilab’s Scintillation Detector Development Center enhanced the University of Nebraska’s future role as a particle detector R&D and production facility for future high energy physics experiments. Overall project objective: This targeted project enhancedmore » the University of Nebraska’s presence in both frontier high energy physics research in DZERO and particle detector development, and it thereby served the goals of the DOE Office of Science and the Experimental Program to Stimulate Competitive Research (EPSCoR) for the state of Nebraska.« less

  10. The Microbiology of Subsurface, Salt-Based Nuclear Waste Repositories: Using Microbial Ecology, Bioenergetics, and Projected Conditions to Help Predict Microbial Effects on Repository Performance

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Swanson, Juliet S.; Cherkouk, Andrea; Arnold, Thuro

    This report summarizes the potential role of microorganisms in salt-based nuclear waste repositories using available information on the microbial ecology of hypersaline environments, the bioenergetics of survival under high ionic strength conditions, and “repository microbiology” related studies. In areas where microbial activity is in question, there may be a need to shift the research focus toward feasibility studies rather than studies that generate actual input for performance assessments. In areas where activity is not necessary to affect performance (e.g., biocolloid transport), repository-relevant data should be generated. Both approaches will lend a realistic perspective to a safety case/performance scenario that willmore » most likely underscore the conservative value of that case.« less

  11. Recent technology products from Space Human Factors research

    NASA Technical Reports Server (NTRS)

    Jenkins, James P.

    1991-01-01

    The goals of the NASA Space Human Factors program and the research carried out concerning human factors are discussed with emphasis given to the development of human performance models, data, and tools. The major products from this program are described, which include the Laser Anthropometric Mapping System; a model of the human body for evaluating the kinematics and dynamics of human motion and strength in microgravity environment; an operational experience data base for verifying and validating the data repository of manned space flights; the Operational Experience Database Taxonomy; and a human-computer interaction laboratory whose products are the display softaware and requirements and the guideline documents and standards for applications on human-computer interaction. Special attention is given to the 'Convoltron', a prototype version of a signal processor for synthesizing the head-related transfer functions.

  12. Multisite Semiautomated Clinical Data Repository for Duplication 15q Syndrome: Study Protocol and Early Uses.

    PubMed

    Ajayi, Oluwaseun Jessica; Smith, Ebony Jeannae; Viangteeravat, Teeradache; Huang, Eunice Y; Nagisetty, Naga Satya V Rao; Urraca, Nora; Lusk, Laina; Finucane, Brenda; Arkilo, Dimitrios; Young, Jennifer; Jeste, Shafali; Thibert, Ronald; Reiter, Lawrence T

    2017-10-18

    Chromosome 15q11.2-q13.1 duplication syndrome (Dup15q syndrome) is a rare disorder caused by duplications of chromosome 15q11.2-q13.1, resulting in a wide range of developmental disabilities in affected individuals. The Dup15q Alliance is an organization that provides family support and promotes research to improve the quality of life of patients living with Dup15q syndrome. Because of the low prevalence of this condition, the establishment of a single research repository would have been difficult and more time consuming without collaboration across multiple institutions. The goal of this project is to establish a national deidentified database with clinical and survey information on individuals diagnosed with Dup15q syndrome. The development of a multiclinic site repository for clinical and survey data on individuals with Dup15q syndrome was initiated and supported by the Dup15q Alliance. Using collaborative workflows, communication protocols, and stakeholder engagement tools, a comprehensive database of patient-centered information was built. We successfully established a self-report populating, centralized repository for Dup15q syndrome research. This repository also resulted in the development of standardized instruments that can be used for other studies relating to developmental disorders. By standardizing the data collection instruments, it allows us integrate our data with other national databases, such as the National Database for Autism Research. A substantial portion of the data collected from the questionnaires was facilitated through direct engagement of participants and their families. This allowed for a more complete set of information to be collected with a minimal turnaround time. We developed a repository that can efficiently be mined for shared clinical phenotypes observed at multiple clinic sites and used as a springboard for future clinical and basic research studies. ©Oluwaseun Jessica Ajayi, Ebony Jeannae Smith, Teeradache Viangteeravat, Eunice Y Huang, Naga Satya V Rao Nagisetty, Nora Urraca, Laina Lusk, Brenda Finucane, Dimitrios Arkilo, Jennifer Young, Shafali Jeste, Ronald Thibert, The Dup15q Alliance, Lawrence T Reiter. Originally published in JMIR Research Protocols (http://www.researchprotocols.org), 18.10.2017.

  13. KUTE-BASE: storing, downloading and exporting MIAME-compliant microarray experiments in minutes rather than hours.

    PubMed

    Draghici, Sorin; Tarca, Adi L; Yu, Longfei; Ethier, Stephen; Romero, Roberto

    2008-03-01

    The BioArray Software Environment (BASE) is a very popular MIAME-compliant, web-based microarray data repository. However in BASE, like in most other microarray data repositories, the experiment annotation and raw data uploading can be very timeconsuming, especially for large microarray experiments. We developed KUTE (Karmanos Universal daTabase for microarray Experiments), as a plug-in for BASE 2.0 that addresses these issues. KUTE provides an automatic experiment annotation feature and a completely redesigned data work-flow that dramatically reduce the human-computer interaction time. For instance, in BASE 2.0 a typical Affymetrix experiment involving 100 arrays required 4 h 30 min of user interaction time forexperiment annotation, and 45 min for data upload/download. In contrast, for the same experiment, KUTE required only 28 min of user interaction time for experiment annotation, and 3.3 min for data upload/download. http://vortex.cs.wayne.edu/kute/index.html.

  14. Centralized mouse repositories.

    PubMed

    Donahue, Leah Rae; Hrabe de Angelis, Martin; Hagn, Michael; Franklin, Craig; Lloyd, K C Kent; Magnuson, Terry; McKerlie, Colin; Nakagata, Naomi; Obata, Yuichi; Read, Stuart; Wurst, Wolfgang; Hörlein, Andreas; Davisson, Muriel T

    2012-10-01

    Because the mouse is used so widely for biomedical research and the number of mouse models being generated is increasing rapidly, centralized repositories are essential if the valuable mouse strains and models that have been developed are to be securely preserved and fully exploited. Ensuring the ongoing availability of these mouse strains preserves the investment made in creating and characterizing them and creates a global resource of enormous value. The establishment of centralized mouse repositories around the world for distributing and archiving these resources has provided critical access to and preservation of these strains. This article describes the common and specialized activities provided by major mouse repositories around the world.

  15. Centralized Mouse Repositories

    PubMed Central

    Donahue, Leah Rae; de Angelis, Martin Hrabe; Hagn, Michael; Franklin, Craig; Lloyd, K. C. Kent; Magnuson, Terry; McKerlie, Colin; Nakagata, Naomi; Obata, Yuichi; Read, Stuart; Wurst, Wolfgang; Hörlein, Andreas; Davisson, Muriel T.

    2013-01-01

    Because the mouse is used so widely for biomedical research and the number of mouse models being generated is increasing rapidly, centralized repositories are essential if the valuable mouse strains and models that have been developed are to be securely preserved and fully exploited. Ensuring the ongoing availability of these mouse strains preserves the investment made in creating and characterizing them and creates a global resource of enormous value. The establishment of centralized mouse repositories around the world for distributing and archiving these resources has provided critical access to and preservation of these strains. This article describes the common and specialized activities provided by major mouse repositories around the world. PMID:22945696

  16. National Date Palm Germplasm Repository

    USDA-ARS?s Scientific Manuscript database

    The National Clonal Germplasm Repository for Citrus & Dates (NCGRCD), located in Riverside, California USA, is a project of the Agricultural Research Service (ARS) of the United States Department of Agriculture (USDA). The NCGRCD maintains a collection of germplasm of date palm (Phoneix dactylifera ...

  17. Two new promising cultivars of mango for Florida

    USDA-ARS?s Scientific Manuscript database

    Mango cultivars are mostly the result of random selections from open pollinated chance seedlings of indigenous or introduced germplasm. The National Germplasm Repository (genebank) at the Subtropical Horticulture Research Station (SHRS) in Miami, Florida is an important mango germplasm repository an...

  18. Experiments, conceptual design, preliminary cost estimates and schedules for an underground research facility

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Korbin, G.; Wollenberg, H.; Wilson, C.

    Plans for an underground research facility are presented, incorporating techniques to assess the hydrological and thermomechanical response of a rock mass to the introduction and long-term isolation of radioactive waste, and to assess the effects of excavation on the hydrologic integrity of a repository and its subsequent backfill, plugging, and sealing. The project is designed to utilize existing mine or civil works for access to experimental areas and is estimated to last 8 years at a total cost for contruction and operation of $39.0 million (1981 dollars). Performing the same experiments in an existing underground research facility would reduce themore » duration to 7-1/2 years and cost $27.7 million as a lower-bound estimate. These preliminary plans and estimates should be revised after specific sites are identified which would accommodate the facility.« less

  19. Building a diabetes screening population data repository using electronic medical records.

    PubMed

    Tuan, Wen-Jan; Sheehy, Ann M; Smith, Maureen A

    2011-05-01

    There has been a rapid advancement of information technology in the area of clinical and population health data management since 2000. However, with the fast growth of electronic medical records (EMRs) and the increasing complexity of information systems, it has become challenging for researchers to effectively access, locate, extract, and analyze information critical to their research. This article introduces an outpatient encounter data framework designed to construct an EMR-based population data repository for diabetes screening research. The outpatient encounter data framework is developed on a hybrid data structure of entity-attribute-value models, dimensional models, and relational models. This design preserves a small number of subject-specific tables essential to key clinical constructs in the data repository. It enables atomic information to be maintained in a transparent and meaningful way to researchers and health care practitioners who need to access data and still achieve the same performance level as conventional data warehouse models. A six-layer information processing strategy is developed to extract and transform EMRs to the research data repository. The data structure also complies with both Health Insurance Portability and Accountability Act regulations and the institutional review board's requirements. Although developed for diabetes screening research, the design of the outpatient encounter data framework is suitable for other types of health service research. It may also provide organizations a tool to improve health care quality and efficiency, consistent with the "meaningful use" objectives of the Health Information Technology for Economic and Clinical Health Act. © 2011 Diabetes Technology Society.

  20. A metadata-driven approach to data repository design.

    PubMed

    Harvey, Matthew J; McLean, Andrew; Rzepa, Henry S

    2017-01-01

    The design and use of a metadata-driven data repository for research data management is described. Metadata is collected automatically during the submission process whenever possible and is registered with DataCite in accordance with their current metadata schema, in exchange for a persistent digital object identifier. Two examples of data preview are illustrated, including the demonstration of a method for integration with commercial software that confers rich domain-specific data analytics without introducing customisation into the repository itself.

  1. Determination of Uncertainties for +III and +IV Actinide Solubilities in the WIPP Geochemistry Model for the 2009 Compliance Recertification Application

    NASA Astrophysics Data System (ADS)

    Ismail, A. E.; Xiong, Y.; Nowak, E. J.; Brush, L. H.

    2009-12-01

    The Waste Isolation Pilot Plant (WIPP) is a U.S. Department of Energy (DOE) repository in southeast New Mexico for defense-related transuranic (TRU) waste. Every five years, the DOE is required to submit an application to the Environmental Protection Agency (EPA) demonstrating the WIPP’s continuing compliance with the applicable EPA regulations governing the repository. Part of this recertification effort involves a performance assessment—a probabilistic evaluation of the repository performance with respect to regulatory limits on the amount of releases from the repository to the accessible environment. One of the models used as part of the performance assessment process is a geochemistry model, which predicts solubilities of the radionuclides in the brines that may enter the repository in the different scenarios considered by the performance assessment. The dissolved actinide source term comprises actinide solubilities, which are input parameters for modeling the transport of radionuclides as a result of brine flow through and from the repository. During a performance assessment, the solubilities are modeled as the product of a “base” solubility determined from calculations based on the chemical conditions expected in the repository, and an uncertainty factor that describes the potential deviations of the model from expected behavior. We will focus here on a discussion of the uncertainties. To compute a cumulative distribution function (CDF) for the uncertainties, we compare published, experimentally measured solubility data to predictions made using the established WIPP geochemistry model. The differences between the solubilities observed for a given experiment and the calculated solubilities from the model are used to form the overall CDF, which is then sampled as part of the performance assessment. We will discuss the methodology used to update the CDF’s for the +III actinides, obtained from data for Nd, Am, and Cm, and the +IV actinides, obtained from data for Th, and present results for the calculations of the updated CDF’s. We compare the CDF’s to the distributions computed for the previous recertification, and discuss the potential impact of the changes on the geochemistry model. This research is funded by WIPP programs administered by the U.S. Department of Energy. Sandia is a multiprogram laboratory operated by Sandia Corporation, a Lockheed Martin Company, for the United States Department of Energy’s National Nuclear Security Administration under contract DE-AC04-94AL85000.

  2. A Repository of Codes of Ethics and Technical Standards in Health Informatics

    PubMed Central

    Zaïane, Osmar R.

    2014-01-01

    We present a searchable repository of codes of ethics and standards in health informatics. It is built using state-of-the-art search algorithms and technologies. The repository will be potentially beneficial for public health practitioners, researchers, and software developers in finding and comparing ethics topics of interest. Public health clinics, clinicians, and researchers can use the repository platform as a one-stop reference for various ethics codes and standards. In addition, the repository interface is built for easy navigation, fast search, and side-by-side comparative reading of documents. Our selection criteria for codes and standards are two-fold; firstly, to maintain intellectual property rights, we index only codes and standards freely available on the internet. Secondly, major international, regional, and national health informatics bodies across the globe are surveyed with the aim of understanding the landscape in this domain. We also look at prevalent technical standards in health informatics from major bodies such as the International Standards Organization (ISO) and the U. S. Food and Drug Administration (FDA). Our repository contains codes of ethics from the International Medical Informatics Association (IMIA), the iHealth Coalition (iHC), the American Health Information Management Association (AHIMA), the Australasian College of Health Informatics (ACHI), the British Computer Society (BCS), and the UK Council for Health Informatics Professions (UKCHIP), with room for adding more in the future. Our major contribution is enhancing the findability of codes and standards related to health informatics ethics by compilation and unified access through the health informatics ethics repository. PMID:25422725

  3. A Structured Approach Using the Systematic Review Data Repository (SRDR): Building the Evidence for Oral Health Interventions in the Population With Intellectual and Developmental Disability.

    PubMed

    Bonardi, Alexandra; Clifford, Christine J; Hadar, Nira

    2016-08-19

    This review describes the methods used for a systematic review of oral health intervention literature in a target population (people with intellectual and developmental disability (I/DD)), which spans a broad range of interventions and study types, conducted with specialized software. The aim of this article is to demonstrate the review strategy, using the free, online systematic review data repository (SRDR) tool, for oral health interventions aimed at reducing disparities between people with I/DD and the general population. Researchers used online title/abstract review (Abstrackr) and data extraction (SRDR) tools to structure the literature review and data extraction. A practicing clinician and an expert methodologist completed the quality review for each study. The data extraction team reported on the experience of using and customizing the SRDR. Using the SRDR, the team developed four extraction templates for eight key questions and completed extraction on 125 articles. This report discusses the advantages and disadvantages of using an electronic tool, such as the SRDR, in completing a systematic review in an area of growing research. This review provides valuable insight for researchers who are considering the use of the SRDR. © The Author(s) 2016.

  4. The Tropical and Subtropical Germplasm Repositories of The National Germplasm System

    USDA-ARS?s Scientific Manuscript database

    Germplasm collections are viewed as a source of genetic diversity to support crop improvement and agricultural research, and germplasm conservation efforts. The United States Department of Agriculture's National Plant Germplasm Repository System (NPGS) is responsible for administering plant genetic ...

  5. PGP repository: a plant phenomics and genomics data publication infrastructure

    PubMed Central

    Arend, Daniel; Junker, Astrid; Scholz, Uwe; Schüler, Danuta; Wylie, Juliane; Lange, Matthias

    2016-01-01

    Plant genomics and phenomics represents the most promising tools for accelerating yield gains and overcoming emerging crop productivity bottlenecks. However, accessing this wealth of plant diversity requires the characterization of this material using state-of-the-art genomic, phenomic and molecular technologies and the release of subsequent research data via a long-term stable, open-access portal. Although several international consortia and public resource centres offer services for plant research data management, valuable digital assets remains unpublished and thus inaccessible to the scientific community. Recently, the Leibniz Institute of Plant Genetics and Crop Plant Research and the German Plant Phenotyping Network have jointly initiated the Plant Genomics and Phenomics Research Data Repository (PGP) as infrastructure to comprehensively publish plant research data. This covers in particular cross-domain datasets that are not being published in central repositories because of its volume or unsupported data scope, like image collections from plant phenotyping and microscopy, unfinished genomes, genotyping data, visualizations of morphological plant models, data from mass spectrometry as well as software and documents. The repository is hosted at Leibniz Institute of Plant Genetics and Crop Plant Research using e!DAL as software infrastructure and a Hierarchical Storage Management System as data archival backend. A novel developed data submission tool was made available for the consortium that features a high level of automation to lower the barriers of data publication. After an internal review process, data are published as citable digital object identifiers and a core set of technical metadata is registered at DataCite. The used e!DAL-embedded Web frontend generates for each dataset a landing page and supports an interactive exploration. PGP is registered as research data repository at BioSharing.org, re3data.org and OpenAIRE as valid EU Horizon 2020 open data archive. Above features, the programmatic interface and the support of standard metadata formats, enable PGP to fulfil the FAIR data principles—findable, accessible, interoperable, reusable. Database URL: http://edal.ipk-gatersleben.de/repos/pgp/ PMID:27087305

  6. Practices of research data curation in institutional repositories: A qualitative view from repository staff

    PubMed Central

    Stvilia, Besiki

    2017-01-01

    The importance of managing research data has been emphasized by the government, funding agencies, and scholarly communities. Increased access to research data increases the impact and efficiency of scientific activities and funding. Thus, many research institutions have established or plan to establish research data curation services as part of their Institutional Repositories (IRs). However, in order to design effective research data curation services in IRs, and to build active research data providers and user communities around those IRs, it is essential to study current data curation practices and provide rich descriptions of the sociotechnical factors and relationships shaping those practices. Based on 13 interviews with 15 IR staff members from 13 large research universities in the United States, this paper provides a rich, qualitative description of research data curation and use practices in IRs. In particular, the paper identifies data curation and use activities in IRs, as well as their structures, roles played, skills needed, contradictions and problems present, solutions sought, and workarounds applied. The paper can inform the development of best practice guides, infrastructure and service templates, as well as education in research data curation in Library and Information Science (LIS) schools. PMID:28301533

  7. Practices of research data curation in institutional repositories: A qualitative view from repository staff.

    PubMed

    Lee, Dong Joon; Stvilia, Besiki

    2017-01-01

    The importance of managing research data has been emphasized by the government, funding agencies, and scholarly communities. Increased access to research data increases the impact and efficiency of scientific activities and funding. Thus, many research institutions have established or plan to establish research data curation services as part of their Institutional Repositories (IRs). However, in order to design effective research data curation services in IRs, and to build active research data providers and user communities around those IRs, it is essential to study current data curation practices and provide rich descriptions of the sociotechnical factors and relationships shaping those practices. Based on 13 interviews with 15 IR staff members from 13 large research universities in the United States, this paper provides a rich, qualitative description of research data curation and use practices in IRs. In particular, the paper identifies data curation and use activities in IRs, as well as their structures, roles played, skills needed, contradictions and problems present, solutions sought, and workarounds applied. The paper can inform the development of best practice guides, infrastructure and service templates, as well as education in research data curation in Library and Information Science (LIS) schools.

  8. Beyond the Repository: A Mixed Method Approach to Providing Access to Collections Online

    ERIC Educational Resources Information Center

    Garrison, Brian Wade

    2013-01-01

    After providing access to over 100 video interviews conducted by a professor with notable entertainers and personalities from film through an institutional repository, an experiment was conducted to discover whether a larger audience could be gained by adding a subset of 32 of these videos to YouTube. The results, over 400,000 views, indicate that…

  9. Applying the institutional review board data repository approach to manage ethical considerations in evaluating and studying medical education

    PubMed Central

    Thayer, Erin K.; Rathkey, Daniel; Miller, Marissa Fuqua; Palmer, Ryan; Mejicano, George C.; Pusic, Martin; Kalet, Adina; Gillespie, Colleen; Carney, Patricia A.

    2016-01-01

    Issue Medical educators and educational researchers continue to improve their processes for managing medical student and program evaluation data using sound ethical principles. This is becoming even more important as curricular innovations are occurring across undergraduate and graduate medical education. Dissemination of findings from this work is critical, and peer-reviewed journals often require an institutional review board (IRB) determination. Approach IRB data repositories, originally designed for the longitudinal study of biological specimens, can be applied to medical education research. The benefits of such an approach include obtaining expedited review for multiple related studies within a single IRB application and allowing for more flexibility when conducting complex longitudinal studies involving large datasets from multiple data sources and/or institutions. In this paper, we inform educators and educational researchers on our analysis of the use of the IRB data repository approach to manage ethical considerations as part of best practices for amassing, pooling, and sharing data for educational research, evaluation, and improvement purposes. Implications Fostering multi-institutional studies while following sound ethical principles in the study of medical education is needed, and the IRB data repository approach has many benefits, especially for longitudinal assessment of complex multi-site data. PMID:27443407

  10. Bioinformatics strategies in life sciences: from data processing and data warehousing to biological knowledge extraction.

    PubMed

    Thiele, Herbert; Glandorf, Jörg; Hufnagel, Peter

    2010-05-27

    With the large variety of Proteomics workflows, as well as the large variety of instruments and data-analysis software available, researchers today face major challenges validating and comparing their Proteomics data. Here we present a new generation of the ProteinScape bioinformatics platform, now enabling researchers to manage Proteomics data from the generation and data warehousing to a central data repository with a strong focus on the improved accuracy, reproducibility and comparability demanded by many researchers in the field. It addresses scientists; current needs in proteomics identification, quantification and validation. But producing large protein lists is not the end point in Proteomics, where one ultimately aims to answer specific questions about the biological condition or disease model of the analyzed sample. In this context, a new tool has been developed at the Spanish Centro Nacional de Biotecnologia Proteomics Facility termed PIKE (Protein information and Knowledge Extractor) that allows researchers to control, filter and access specific information from genomics and proteomic databases, to understand the role and relationships of the proteins identified in the experiments. Additionally, an EU funded project, ProDac, has coordinated systematic data collection in public standards-compliant repositories like PRIDE. This will cover all aspects from generating MS data in the laboratory, assembling the whole annotation information and storing it together with identifications in a standardised format.

  11. Demonstrating the Open Data Repository's Data Publisher: The CheMin Database

    NASA Astrophysics Data System (ADS)

    Stone, N.; Lafuente, B.; Bristow, T.; Pires, A.; Keller, R. M.; Downs, R. T.; Blake, D.; Dateo, C. E.; Fonda, M.

    2018-04-01

    The Open Data Repository's Data Publisher aims to provide an easy-to-use software tool that will allow researchers to create and publish database templates and related data. The CheMin Database developed using this framework is shown as an example.

  12. A laboratory validation study of the time-lapse oscillatory pumping test for leakage detection in geological repositories

    NASA Astrophysics Data System (ADS)

    Sun, Alexander Y.; Lu, Jiemin; Islam, Akand

    2017-05-01

    Geologic repositories are extensively used for disposing byproducts in mineral and energy industries. The safety and reliability of these repositories are a primary concern to environmental regulators and the public. Time-lapse oscillatory pumping test (OPT) has been introduced recently as a pressure-based technique for detecting potential leakage in geologic repositories. By routinely conducting OPT at a number of pulsing frequencies, an operator may identify the potential repository anomalies in the frequency domain, alleviating the ambiguity caused by reservoir noise and improving the signal-to-noise ratio. Building on previous theoretical and field studies, this work performed a series of laboratory experiments to validate the concept of time-lapse OPT using a custom made, stainless steel tank under relatively high pressures. The experimental configuration simulates a miniature geologic storage repository consisting of three layers (i.e., injection zone, caprock, and above-zone aquifer). Results show that leakage in the injection zone led to deviations in the power spectrum of observed pressure data, and the amplitude of which also increases with decreasing pulsing frequencies. The experimental results are further analyzed by developing a 3D flow model, using which the model parameters are estimated through frequency domain inversion.

  13. Scientific information repository assisting reflectance spectrometry in legal medicine.

    PubMed

    Belenki, Liudmila; Sterzik, Vera; Bohnert, Michael; Zimmermann, Klaus; Liehr, Andreas W

    2012-06-01

    Reflectance spectrometry is a fast and reliable method for the characterization of human skin if the spectra are analyzed with respect to a physical model describing the optical properties of human skin. For a field study performed at the Institute of Legal Medicine and the Freiburg Materials Research Center of the University of Freiburg, a scientific information repository has been developed, which is a variant of an electronic laboratory notebook and assists in the acquisition, management, and high-throughput analysis of reflectance spectra in heterogeneous research environments. At the core of the repository is a database management system hosting the master data. It is filled with primary data via a graphical user interface (GUI) programmed in Java, which also enables the user to browse the database and access the results of data analysis. The latter is carried out via Matlab, Python, and C programs, which retrieve the primary data from the scientific information repository, perform the analysis, and store the results in the database for further usage.

  14. Gas and water flow in an excavation-induced fracture network around an underground drift: A case study for a radioactive waste repository in clay rock

    NASA Astrophysics Data System (ADS)

    de La Vaissière, Rémi; Armand, Gilles; Talandier, Jean

    2015-02-01

    The Excavation Damaged Zone (EDZ) surrounding a drift, and in particular its evolution, is being studied for the performance assessment of a radioactive waste underground repository. A specific experiment (called CDZ) was designed and implemented in the Meuse/Haute-Marne Underground Research Laboratory (URL) in France to investigate the EDZ. This experiment is dedicated to study the evolution of the EDZ hydrogeological properties (conductivity and specific storage) of the Callovo-Oxfordian claystone under mechanical compression and artificial hydration. Firstly, a loading cycle applied on a drift wall was performed to simulate the compression effect from bentonite swelling in a repository drift (bentonite is a clay material to be used to seal drifts and shafts for repository closure purpose). Gas tests (permeability tests with nitrogen and tracer tests with helium) were conducted during the first phase of the experiment. The results showed that the fracture network within the EDZ was initially interconnected and opened for gas flow (particularly along the drift) and then progressively closed with the increasing mechanical stress applied on the drift wall. Moreover, the evolution of the EDZ after unloading indicated a self-sealing process. Secondly, the remaining fracture network was resaturated to demonstrate the ability to self-seal of the COx claystone without mechanical loading by conducting from 11 to 15 repetitive hydraulic tests with monitoring of the hydraulic parameters. During this hydration process, the EDZ effective transmissivity dropped due to the swelling of the clay materials near the fracture network. The hydraulic conductivity evolution was relatively fast during the first few days. Low conductivities ranging at 10-10 m/s were observed after four months. Conversely, the specific storage showed an erratic evolution during the first phase of hydration (up to 60 days). Some uncertainty remains on this parameter due to volumetric strain during the sealing of the fractures. The hydration was stopped after one year and cross-hole hydraulic tests were performed to determine more accurately the specific storage as well as the hydraulic conductivity at a meter-scale. All hydraulic conductivity values measured at the injection interval and at the observation intervals were all below 10-10 m/s. Moreover, the preferential inter-connectivity along the drift disappeared. Specific storage values at the observation and injection intervals were similar. Furthermore they were in agreement with the value obtained at the injection interval within the second hydration phase (60 days after starting hydration). The graphical abstract synthesizes the evolution of the hydraulic/gas conductivity for 8 intervals since the beginning of the CDZ experiment. The conductivity limit of 10-10 m/s corresponds to the lower bound hydraulic definition of the EDZ and it is demonstrated that EDZ can be sealed. This is a significant result in the demonstration of the long-term safety of a repository.

  15. Proteomics data repositories

    PubMed Central

    Riffle, Michael; Eng, Jimmy K.

    2010-01-01

    The field of proteomics, particularly the application of mass spectrometry analysis to protein samples, is well-established and growing rapidly. Proteomics studies generate large volumes of raw experimental data and inferred biological results. To facilitate the dissemination of these data, centralized data repositories have been developed that make the data and results accessible to proteomics researchers and biologists alike. This review of proteomics data repositories focuses exclusively on freely-available, centralized data resources that disseminate or store experimental mass spectrometry data and results. The resources chosen reflect a current “snapshot” of the state of resources available with an emphasis placed on resources that may be of particular interest to yeast researchers. Resources are described in terms of their intended purpose and the features and functionality provided to users. PMID:19795424

  16. Feasibility of an International Multiple Sclerosis Rehabilitation Data Repository

    PubMed Central

    Bradford, Elissa Held; Baert, Ilse; Finlayson, Marcia; Feys, Peter

    2018-01-01

    Abstract Background: Multiple sclerosis (MS) rehabilitation evidence is limited due to methodological factors, which may be addressed by a data repository. We describe the perceived challenges of, motivators for, interest in participating in, and key features of an international MS rehabilitation data repository. Methods: A multimethod sequential investigation was performed with the results of two focus groups, using nominal group technique, and study aims informing the development of an online questionnaire. Percentage agreement and key quotations illustrated questionnaire findings. Subgroup comparisons were made between clinicians and researchers and between participants in North America and Europe. Results: Rehabilitation professionals from 25 countries participated (focus groups: n = 21; questionnaire: n = 166). The top ten challenges (C) and motivators (M) identified by the focus groups were database control/management (C); ethical/legal concerns (C); data quality (C); time, effort, and cost (C); best practice (M); uniformity (C); sustainability (C); deeper analysis (M); collaboration (M); and identifying research needs (M). Percentage agreement with questionnaire statements regarding challenges to, motivators for, interest in, and key features of a successful repository was at least 80%, 85%, 72%, and 83%, respectively, across each group of statements. Questionnaire subgroup analysis revealed a few differences (P < .05), including that clinicians more strongly identified with improving best practice as a motivator. Conclusions: Findings support clinician and researcher interest in and potential for success of an international MS rehabilitation data repository if prioritized challenges and motivators are addressed and key features are included. PMID:29507539

  17. Feasibility of an International Multiple Sclerosis Rehabilitation Data Repository: Perceived Challenges and Motivators for Sharing Data.

    PubMed

    Bradford, Elissa Held; Baert, Ilse; Finlayson, Marcia; Feys, Peter; Wagner, Joanne

    2018-01-01

    Multiple sclerosis (MS) rehabilitation evidence is limited due to methodological factors, which may be addressed by a data repository. We describe the perceived challenges of, motivators for, interest in participating in, and key features of an international MS rehabilitation data repository. A multimethod sequential investigation was performed with the results of two focus groups, using nominal group technique, and study aims informing the development of an online questionnaire. Percentage agreement and key quotations illustrated questionnaire findings. Subgroup comparisons were made between clinicians and researchers and between participants in North America and Europe. Rehabilitation professionals from 25 countries participated (focus groups: n = 21; questionnaire: n = 166). The top ten challenges (C) and motivators (M) identified by the focus groups were database control/management (C); ethical/legal concerns (C); data quality (C); time, effort, and cost (C); best practice (M); uniformity (C); sustainability (C); deeper analysis (M); collaboration (M); and identifying research needs (M). Percentage agreement with questionnaire statements regarding challenges to, motivators for, interest in, and key features of a successful repository was at least 80%, 85%, 72%, and 83%, respectively, across each group of statements. Questionnaire subgroup analysis revealed a few differences (P < .05), including that clinicians more strongly identified with improving best practice as a motivator. Findings support clinician and researcher interest in and potential for success of an international MS rehabilitation data repository if prioritized challenges and motivators are addressed and key features are included.

  18. The Experiment Factory: Standardizing Behavioral Experiments.

    PubMed

    Sochat, Vanessa V; Eisenberg, Ian W; Enkavi, A Zeynep; Li, Jamie; Bissett, Patrick G; Poldrack, Russell A

    2016-01-01

    The administration of behavioral and experimental paradigms for psychology research is hindered by lack of a coordinated effort to develop and deploy standardized paradigms. While several frameworks (Mason and Suri, 2011; McDonnell et al., 2012; de Leeuw, 2015; Lange et al., 2015) have provided infrastructure and methods for individual research groups to develop paradigms, missing is a coordinated effort to develop paradigms linked with a system to easily deploy them. This disorganization leads to redundancy in development, divergent implementations of conceptually identical tasks, disorganized and error-prone code lacking documentation, and difficulty in replication. The ongoing reproducibility crisis in psychology and neuroscience research (Baker, 2015; Open Science Collaboration, 2015) highlights the urgency of this challenge: reproducible research in behavioral psychology is conditional on deployment of equivalent experiments. A large, accessible repository of experiments for researchers to develop collaboratively is most efficiently accomplished through an open source framework. Here we present the Experiment Factory, an open source framework for the development and deployment of web-based experiments. The modular infrastructure includes experiments, virtual machines for local or cloud deployment, and an application to drive these components and provide developers with functions and tools for further extension. We release this infrastructure with a deployment (http://www.expfactory.org) that researchers are currently using to run a set of over 80 standardized web-based experiments on Amazon Mechanical Turk. By providing open source tools for both deployment and development, this novel infrastructure holds promise to bring reproducibility to the administration of experiments, and accelerate scientific progress by providing a shared community resource of psychological paradigms.

  19. The Experiment Factory: Standardizing Behavioral Experiments

    PubMed Central

    Sochat, Vanessa V.; Eisenberg, Ian W.; Enkavi, A. Zeynep; Li, Jamie; Bissett, Patrick G.; Poldrack, Russell A.

    2016-01-01

    The administration of behavioral and experimental paradigms for psychology research is hindered by lack of a coordinated effort to develop and deploy standardized paradigms. While several frameworks (Mason and Suri, 2011; McDonnell et al., 2012; de Leeuw, 2015; Lange et al., 2015) have provided infrastructure and methods for individual research groups to develop paradigms, missing is a coordinated effort to develop paradigms linked with a system to easily deploy them. This disorganization leads to redundancy in development, divergent implementations of conceptually identical tasks, disorganized and error-prone code lacking documentation, and difficulty in replication. The ongoing reproducibility crisis in psychology and neuroscience research (Baker, 2015; Open Science Collaboration, 2015) highlights the urgency of this challenge: reproducible research in behavioral psychology is conditional on deployment of equivalent experiments. A large, accessible repository of experiments for researchers to develop collaboratively is most efficiently accomplished through an open source framework. Here we present the Experiment Factory, an open source framework for the development and deployment of web-based experiments. The modular infrastructure includes experiments, virtual machines for local or cloud deployment, and an application to drive these components and provide developers with functions and tools for further extension. We release this infrastructure with a deployment (http://www.expfactory.org) that researchers are currently using to run a set of over 80 standardized web-based experiments on Amazon Mechanical Turk. By providing open source tools for both deployment and development, this novel infrastructure holds promise to bring reproducibility to the administration of experiments, and accelerate scientific progress by providing a shared community resource of psychological paradigms. PMID:27199843

  20. Examining Data Repository Guidelines for Qualitative Data Sharing.

    PubMed

    Antes, Alison L; Walsh, Heidi A; Strait, Michelle; Hudson-Vitale, Cynthia R; DuBois, James M

    2018-02-01

    Qualitative data provide rich information on research questions in diverse fields. Recent calls for increased transparency and openness in research emphasize data sharing. However, qualitative data sharing has yet to become the norm internationally and is particularly uncommon in the United States. Guidance for archiving and secondary use of qualitative data is required for progress in this regard. In this study, we review the benefits and concerns associated with qualitative data sharing and then describe the results of a content analysis of guidelines from international repositories that archive qualitative data. A minority of repositories provide qualitative data sharing guidelines. Of the guidelines available, there is substantial variation in whether specific topics are addressed. Some topics, such as removing direct identifiers, are consistently addressed, while others, such as providing an anonymization log, are not. We discuss the implications of our study for education, best practices, and future research.

  1. De-identification of Medical Images with Retention of Scientific Research Value

    PubMed Central

    Maffitt, David R.; Smith, Kirk E.; Kirby, Justin S.; Clark, Kenneth W.; Freymann, John B.; Vendt, Bruce A.; Tarbox, Lawrence R.; Prior, Fred W.

    2015-01-01

    Online public repositories for sharing research data allow investigators to validate existing research or perform secondary research without the expense of collecting new data. Patient data made publicly available through such repositories may constitute a breach of personally identifiable information if not properly de-identified. Imaging data are especially at risk because some intricacies of the Digital Imaging and Communications in Medicine (DICOM) format are not widely understood by researchers. If imaging data still containing protected health information (PHI) were released through a public repository, a number of different parties could be held liable, including the original researcher who collected and submitted the data, the original researcher’s institution, and the organization managing the repository. To minimize these risks through proper de-identification of image data, one must understand what PHI exists and where that PHI resides, and one must have the tools to remove PHI without compromising the scientific integrity of the data. DICOM public elements are defined by the DICOM Standard. Modality vendors use private elements to encode acquisition parameters that are not yet defined by the DICOM Standard, or the vendor may not have updated an existing software product after DICOM defined new public elements. Because private elements are not standardized, a common de-identification practice is to delete all private elements, removing scientifically useful data as well as PHI. Researchers and publishers of imaging data can use the tools and process described in this article to de-identify DICOM images according to current best practices. ©RSNA, 2015 PMID:25969931

  2. e!DAL - a framework to store, share and publish research data

    PubMed Central

    2014-01-01

    Background The life-science community faces a major challenge in handling “big data”, highlighting the need for high quality infrastructures capable of sharing and publishing research data. Data preservation, analysis, and publication are the three pillars in the “big data life cycle”. The infrastructures currently available for managing and publishing data are often designed to meet domain-specific or project-specific requirements, resulting in the repeated development of proprietary solutions and lower quality data publication and preservation overall. Results e!DAL is a lightweight software framework for publishing and sharing research data. Its main features are version tracking, metadata management, information retrieval, registration of persistent identifiers (DOI), an embedded HTTP(S) server for public data access, access as a network file system, and a scalable storage backend. e!DAL is available as an API for local non-shared storage and as a remote API featuring distributed applications. It can be deployed “out-of-the-box” as an on-site repository. Conclusions e!DAL was developed based on experiences coming from decades of research data management at the Leibniz Institute of Plant Genetics and Crop Plant Research (IPK). Initially developed as a data publication and documentation infrastructure for the IPK’s role as a data center in the DataCite consortium, e!DAL has grown towards being a general data archiving and publication infrastructure. The e!DAL software has been deployed into the Maven Central Repository. Documentation and Software are also available at: http://edal.ipk-gatersleben.de. PMID:24958009

  3. e!DAL--a framework to store, share and publish research data.

    PubMed

    Arend, Daniel; Lange, Matthias; Chen, Jinbo; Colmsee, Christian; Flemming, Steffen; Hecht, Denny; Scholz, Uwe

    2014-06-24

    The life-science community faces a major challenge in handling "big data", highlighting the need for high quality infrastructures capable of sharing and publishing research data. Data preservation, analysis, and publication are the three pillars in the "big data life cycle". The infrastructures currently available for managing and publishing data are often designed to meet domain-specific or project-specific requirements, resulting in the repeated development of proprietary solutions and lower quality data publication and preservation overall. e!DAL is a lightweight software framework for publishing and sharing research data. Its main features are version tracking, metadata management, information retrieval, registration of persistent identifiers (DOI), an embedded HTTP(S) server for public data access, access as a network file system, and a scalable storage backend. e!DAL is available as an API for local non-shared storage and as a remote API featuring distributed applications. It can be deployed "out-of-the-box" as an on-site repository. e!DAL was developed based on experiences coming from decades of research data management at the Leibniz Institute of Plant Genetics and Crop Plant Research (IPK). Initially developed as a data publication and documentation infrastructure for the IPK's role as a data center in the DataCite consortium, e!DAL has grown towards being a general data archiving and publication infrastructure. The e!DAL software has been deployed into the Maven Central Repository. Documentation and Software are also available at: http://edal.ipk-gatersleben.de.

  4. An ontology based information system for the management of institutional repository's collections

    NASA Astrophysics Data System (ADS)

    Tsolakidis, A.; Kakoulidis, P.; Skourlas, C.

    2015-02-01

    In this paper we discuss a simple methodological approach to create, and customize institutional repositories for the domain of the technological education. The use of the open source software platform of DSpace is proposed to build up the repository application and provide access to digital resources including research papers, dissertations, administrative documents, educational material, etc. Also the use of owl ontologies is proposed for indexing and accessing the various, heterogeneous items stored in the repository. Customization and operation of a platform for the selection and use of terms or parts of similar existing owl ontologies is also described. This platform could be based on the open source software Protégé that supports owl, is widely used, and also supports visualization, SPARQL etc. The combined use of the owl platform and the DSpace repository form a basis for creating customized ontologies, accommodating the semantic metadata of items and facilitating searching.

  5. NUTRITION w/Repository

    NASA Image and Video Library

    2009-06-06

    ISS020-E-007566 (6 June 2009) --- European Space Agency astronaut Frank De Winne, Expedition 20 flight engineer, prepares to put samples in the Minus Eighty Laboratory Freezer for ISS (MELFI) in the Kibo laboratory of the International Space Station. Samples were taken as part of the Nutritional Status Assessment (Nutrition) with Repository experiment, a study done by NASA to date of human physiologic changes during long-duration spaceflight.

  6. Models for Evaluating and Improving Architecture Competence

    DTIC Science & Technology

    2008-03-01

    learned better methods than it engaged in the past. 36 | CMU/SEI-2008-TR-006 SOFTWARE ENGINEERING INSTITUTE | 37 6 Considering the Models ...and groups must have a repository of ac- cumulated knowledge and experience. The Organizational Learning model provides a way to eva- luate how...effective that repository is. It also tells us how ―mindful‖ the learning needs to be. The organizational coordination model

  7. Urban History in 4 Dimensions - Supporting Research and Education

    NASA Astrophysics Data System (ADS)

    Münster, S.; Friedrichs, K.; Kröber, C.; Bruschke, J.; Henze, F.; Maiwald, F.; Niebling, F.

    2017-08-01

    The new research group on the four-dimensional research and communication of urban history (Urban History 4D) aims to investigate and develop methods and technologies to access extensive repositories of historical media and their contextual information in a spatial model, with an additional temporal component. This will make content accessible to different target groups, researchers and the public, via a 4D browser. A location-dependent augmented-reality representation can be used as an information base, research tool, and means of communicating historical knowledge. The data resources for this research include extensive holdings of historical photographs of Dresden, which have documented the city over the decades, and digitized map collections from the Deutsche Fotothek (German photographic collection) platform. These will lay the foundation for a prototype model which will give users a virtual experience of historic parts of Dresden.

  8. Preservation Health Check: Monitoring Threats to Digital Repository Content

    ERIC Educational Resources Information Center

    Kool, Wouter; van der Werf, Titia; Lavoie, Brian

    2014-01-01

    The Preservation Health Check (PHC) project, undertaken as a joint effort by Open Planets Foundation (OPF) and OCLC Research, aims to evaluate the usefulness of the preservation metadata created and maintained by operational repositories for assessing basic preservation properties. The PHC project seeks to develop an implementable logic to support…

  9. Embracing the Future: Embedding Digital Repositories in the University of London. Technical Report

    ERIC Educational Resources Information Center

    Hoorens, Stijn; van Dijk, Lidia Villalba; van Stolk, Christian

    2008-01-01

    Digital repositories can help Higher Education Institutions (HEIs) to develop coherent and coordinated approaches to capture, identify, store and retrieve intellectual assets such as datasets, course material and research papers. With the advances of technology, an increasing number of Higher Education Institutions are implementing digital…

  10. Collaborative Posters Develop Students' Ability to Communicate about Undervalued Scientific Resources to Nonscientists.

    PubMed

    Mayfield, Teresa J; Olimpo, Jeffrey T; Floyd, Kevin W; Greenbaum, Eli

    2018-01-01

    Scientists are increasingly called upon to communicate with the public, yet most never receive formal training in this area. Public understanding is particularly critical to maintaining support for undervalued resources such as biological collections, research data repositories, and expensive equipment. We describe activities carried out in an inquiry-driven organismal biology laboratory course designed to engage a diverse student body using biological collections. The goals of this cooperative learning experience were to increase students' ability to locate and comprehend primary research articles, and to communicate the importance of an undervalued scientific resource to nonscientists. Our results indicate that collaboratively created, research-focused informational posters are an effective tool for achieving these goals and may be applied in other disciplines or classroom settings.

  11. Interoperability Across the Stewardship Spectrum in the DataONE Repository Federation

    NASA Astrophysics Data System (ADS)

    Jones, M. B.; Vieglais, D.; Wilson, B. E.

    2016-12-01

    Thousands of earth and environmental science repositories serve many researchers and communities, each with their own community and legal mandates, sustainability models, and historical infrastructure. These repositories span the stewardship spectrum from highly curated collections that employ large numbers of staff members to review and improve data, to small, minimal budget repositories that accept data caveat emptor and where all responsibility for quality lies with the submitter. Each repository fills a niche, providing services that meet the stewardship tradeoffs of one or more communities. We have reviewed these stewardship tradeoffs for several DataONE member repositories ranging from minimally (KNB) to highly curated (Arctic Data Center), as well as general purpose (Dryad) to highly discipline or project specific (NEON). The rationale behind different levels of stewardship reflect resolution of these tradeoffs. Some repositories aim to encourage extensive uptake by keeping processes simple and minimizing the amount of information collected, but this limits the long-term utility of the data and the search, discovery, and integration systems that are possible. Other repositories require extensive metadata input, review, and assessment, allowing for excellent preservation, discovery, and integration but at the cost of significant time for submitters and expense for curatorial staff. DataONE recognizes these different levels of curation, and attempts to embrace them to create a federation that is useful across the stewardship spectrum. DataONE provides a tiered model for repositories with growing utility of DataONE services at higher tiers of curation. The lowest tier supports read-only access to data and requires little more than title and contact metadata. Repositories can gradually phase in support for higher levels of metadata and services as needed. These tiered capabilities are possible through flexible support for multiple metadata standards and services, where repositories can incrementally increase their requirements as they want to satisfy more use cases. Within DataONE, metadata search services support minimal metadata models, but significantly expanded precision and recall become possible when repositories provide more extensively curated metadata.

  12. Optimizing Resources for Trustworthiness and Scientific Impact of Domain Repositories

    NASA Astrophysics Data System (ADS)

    Lehnert, K.

    2017-12-01

    Domain repositories, i.e. data archives tied to specific scientific communities, are widely recognized and trusted by their user communities for ensuring a high level of data quality, enhancing data value, access, and reuse through a unique combination of disciplinary and digital curation expertise. Their data services are guided by the practices and values of the specific community they serve and designed to support the advancement of their science. Domain repositories need to meet user expectations for scientific utility in order to be successful, but they also need to fulfill the requirements for trustworthy repository services to be acknowledged by scientists, funders, and publishers as a reliable facility that curates and preserves data following international standards. Domain repositories therefore need to carefully plan and balance investments to optimize the scientific impact of their data services and user satisfaction on the one hand, while maintaining a reliable and robust operation of the repository infrastructure on the other hand. Staying abreast of evolving repository standards to certify as a trustworthy repository and conducting a regular self-assessment and certification alone requires resources that compete with the demands for improving data holdings or usability of systems. The Interdisciplinary Earth Data Alliance (IEDA), a data facility funded by the US National Science Foundation, operates repositories for geochemical, marine Geoscience, and Antarctic research data, while also maintaining data products (global syntheses) and data visualization and analysis tools that are of high value for the science community and have demonstrated considerable scientific impact. Balancing the investments in the growth and utility of the syntheses with resources required for certifcation of IEDA's repository services has been challenging, and a major self-assessment effort has been difficult to accommodate. IEDA is exploring a partnership model to share generic repository functions (e.g. metadata registration, long-term archiving) with other repositories. This could substantially reduce the effort of certification and allow effort to focus on the domain-specific data curation and value-added services.

  13. Preliminary safety analysis of the Baita Bihor radioactive waste repository, Romania

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Little, Richard; Bond, Alex; Watson, Sarah

    2007-07-01

    A project funded under the European Commission's Phare Programme 2002 has undertaken an in-depth analysis of the operational and post-closure safety of the Baita Bihor repository. The repository has accepted low- and some intermediate-level radioactive waste from industry, medical establishments and research activities since 1985 and the current estimate is that disposals might continue for around another 20 to 35 years. The analysis of the operational and post-closure safety of the Baita Bihor repository was carried out in two iterations, with the second iteration resulting in reduced uncertainties, largely as a result taking into account new information on the hydrologymore » and hydrogeology of the area, collected as part of the project. Impacts were evaluated for the maximum potential inventory that might be available for disposal to Baita Bihor for a number of operational and postclosure scenarios and associated conceptual models. The results showed that calculated impacts were below the relevant regulatory criteria. In light of the assessment, a number of recommendations relating to repository operation, optimisation of repository engineering and waste disposals, and environmental monitoring were made. (authors)« less

  14. The MIMIC Code Repository: enabling reproducibility in critical care research.

    PubMed

    Johnson, Alistair Ew; Stone, David J; Celi, Leo A; Pollard, Tom J

    2018-01-01

    Lack of reproducibility in medical studies is a barrier to the generation of a robust knowledge base to support clinical decision-making. In this paper we outline the Medical Information Mart for Intensive Care (MIMIC) Code Repository, a centralized code base for generating reproducible studies on an openly available critical care dataset. Code is provided to load the data into a relational structure, create extractions of the data, and reproduce entire analysis plans including research studies. Concepts extracted include severity of illness scores, comorbid status, administrative definitions of sepsis, physiologic criteria for sepsis, organ failure scores, treatment administration, and more. Executable documents are used for tutorials and reproduce published studies end-to-end, providing a template for future researchers to replicate. The repository's issue tracker enables community discussion about the data and concepts, allowing users to collaboratively improve the resource. The centralized repository provides a platform for users of the data to interact directly with the data generators, facilitating greater understanding of the data. It also provides a location for the community to collaborate on necessary concepts for research progress and share them with a larger audience. Consistent application of the same code for underlying concepts is a key step in ensuring that research studies on the MIMIC database are comparable and reproducible. By providing open source code alongside the freely accessible MIMIC-III database, we enable end-to-end reproducible analysis of electronic health records. © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association.

  15. DSpace and customized controlled vocabularies

    NASA Astrophysics Data System (ADS)

    Skourlas, C.; Tsolakidis, A.; Kakoulidis, P.; Giannakopoulos, G.

    2015-02-01

    The open source platform of DSpace could be defined as a repository application used to provide access to digital resources. DSpace is installed and used by more than 1000 organizations worldwide. A predefined taxonomy of keyword, called the Controlled Vocabulary, can be used for describing and accessing the information items stored in the repository. In this paper, we describe how the users can create, and customize their own vocabularies. Various heterogeneous items, such as research papers, videos, articles and educational material of the repository, can be indexed in order to provide advanced search functionality using new controlled vocabularies.

  16. Iterative performance assessments as a regulatory tool for evaluating repository safety: How experiences from SKI Project-90 were used in formulating the new performance assessment project SITE-94

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Andersson, J.

    1993-12-31

    The Swedish Nuclear Power Inspectorate, SKI, regulatory research program has to prepare for the process of licensing a repository for spent nuclear fuel, by building up the necessary knowledge and review capacity. SKIs main strategy for meeting this demand is to develop an independent performance assessment capability. SKIs first own performance assessment project, Project-90, was completed in 1991 and is now followed by a new project, SITE-94. SITE-94 is based on conclusions reached within Project-90. An independent review of Project-90, carried out by a NEA team of experts, has also contributed to the formation of the project. Another important reasonmore » for the project is that the implementing organization in Sweden, SKB, has proposed to submit an application to start detailed investigation of a repository candidate site around 1997. SITE-94 is a performance assessment of a hypothetical repository at a real site. The main objective of the project is to determine how site specific data should be assimilated into the performance assessment process, and to evaluate how uncertainties inherent in site characterization will influence performance assessment results. This will be addressed by exploring multiple interpretations, conceptual models, and parameters consistent with the site data. The site evaluation will strive for consistency between geological, hydrological, rock mechanical, and geochemical descriptions. Other important elements of SITE-94 are the development of a practical and defensible methodology for defining, constructing and analyzing scenarios, the development of approaches for treatment of uncertainties, evaluation of canister integrity, and the development and application of an appropriate quality assurance plan for performance assessments.« less

  17. A scoping review of online repositories of quality improvement projects, interventions and initiatives in healthcare.

    PubMed

    Bytautas, Jessica P; Gheihman, Galina; Dobrow, Mark J

    2017-04-01

    Quality improvement (QI) is becoming an important focal point for health systems. There is increasing interest among health system stakeholders to learn from and share experiences on the use of QI methods and approaches in their work. Yet there are few easily accessible, online repositories dedicated to documenting QI activity. We conducted a scoping review of publicly available, web-based QI repositories to (i) identify current approaches to sharing information on QI practices; (ii) categorise these approaches based on hosting, scope and size, content acquisition and eligibility, content format and search, and evaluation and engagement characteristics; and (iii) review evaluations of the design, usefulness and impact of their online QI practice repositories. The search strategy consisted of traditional database and grey literature searches, as well as expert consultation, with the ultimate aim of identifying and describing QI repositories of practices undertaken in a healthcare context. We identified 13 QI repositories and found substantial variation across the five categories. The QI repositories used different terminology (eg, practices vs case studies) and approaches to content acquisition, and varied in terms of primary areas of focus. All provided some means for organising content according to categories or themes and most provided at least rudimentary keyword search functionality. Notably, none of the QI repositories included evaluations of their impact. With growing interest in sharing and spreading best practices and increasing reliance on QI as a key contributor to health system performance, the role of QI repositories is likely to expand. Designing future QI repositories based on knowledge of the range and type of features available is an important starting point for improving their usefulness and impact. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.

  18. Semantic Web repositories for genomics data using the eXframe platform.

    PubMed

    Merrill, Emily; Corlosquet, Stéphane; Ciccarese, Paolo; Clark, Tim; Das, Sudeshna

    2014-01-01

    With the advent of inexpensive assay technologies, there has been an unprecedented growth in genomics data as well as the number of databases in which it is stored. In these databases, sample annotation using ontologies and controlled vocabularies is becoming more common. However, the annotation is rarely available as Linked Data, in a machine-readable format, or for standardized queries using SPARQL. This makes large-scale reuse, or integration with other knowledge bases very difficult. To address this challenge, we have developed the second generation of our eXframe platform, a reusable framework for creating online repositories of genomics experiments. This second generation model now publishes Semantic Web data. To accomplish this, we created an experiment model that covers provenance, citations, external links, assays, biomaterials used in the experiment, and the data collected during the process. The elements of our model are mapped to classes and properties from various established biomedical ontologies. Resource Description Framework (RDF) data is automatically produced using these mappings and indexed in an RDF store with a built-in Sparql Protocol and RDF Query Language (SPARQL) endpoint. Using the open-source eXframe software, institutions and laboratories can create Semantic Web repositories of their experiments, integrate it with heterogeneous resources and make it interoperable with the vast Semantic Web of biomedical knowledge.

  19. NUTRITION w/Repository

    NASA Image and Video Library

    2009-06-06

    ISS020-E-007577 (6 June 2009) --- European Space Agency astronaut Frank De Winne, Expedition 20 flight engineer, returns a dewar tray to the Minus Eighty Laboratory Freezer for ISS (MELFI) after inserting biological samples into the trays in the Kibo laboratory of the International Space Station. Samples were taken as part of the Nutritional Status Assessment (Nutrition) with Repository experiment, a study done by NASA to date of human physiologic changes during long-duration spaceflight.

  20. Development of anomaly detection models for deep subsurface monitoring

    NASA Astrophysics Data System (ADS)

    Sun, A. Y.

    2017-12-01

    Deep subsurface repositories are used for waste disposal and carbon sequestration. Monitoring deep subsurface repositories for potential anomalies is challenging, not only because the number of sensor networks and the quality of data are often limited, but also because of the lack of labeled data needed to train and validate machine learning (ML) algorithms. Although physical simulation models may be applied to predict anomalies (or the system's nominal state for that sake), the accuracy of such predictions may be limited by inherent conceptual and parameter uncertainties. The main objective of this study was to demonstrate the potential of data-driven models for leakage detection in carbon sequestration repositories. Monitoring data collected during an artificial CO2 release test at a carbon sequestration repository were used, which include both scalar time series (pressure) and vector time series (distributed temperature sensing). For each type of data, separate online anomaly detection algorithms were developed using the baseline experiment data (no leak) and then tested on the leak experiment data. Performance of a number of different online algorithms was compared. Results show the importance of including contextual information in the dataset to mitigate the impact of reservoir noise and reduce false positive rate. The developed algorithms were integrated into a generic Web-based platform for real-time anomaly detection.

  1. Modelling geochemical and microbial consumption of dissolved oxygen after backfilling a high level radiactive waste repository.

    PubMed

    Yang, Changbing; Samper, Javier; Molinero, Jorge; Bonilla, Mercedes

    2007-08-15

    Dissolved oxygen (DO) left in the voids of buffer and backfill materials of a deep geological high level radioactive waste (HLW) repository could cause canister corrosion. Available data from laboratory and in situ experiments indicate that microbes play a substantial role in controlling redox conditions near a HLW repository. This paper presents the application of a coupled hydro-bio-geochemical model to evaluate geochemical and microbial consumption of DO in bentonite porewater after backfilling of a HLW repository designed according to the Swedish reference concept. In addition to geochemical reactions, the model accounts for dissolved organic carbon (DOC) respiration and methane oxidation. Parameters for microbial processes were derived from calibration of the REX in situ experiment carried out at the Aspö underground laboratory. The role of geochemical and microbial processes in consuming DO is evaluated for several scenarios. Numerical results show that both geochemical and microbial processes are relevant for DO consumption. However, the time needed to consume the DO trapped in the bentonite buffer decreases dramatically from several hundreds of years when only geochemical processes are considered to a few weeks when both geochemical reactions and microbially-mediated DOC respiration and methane oxidation are taken into account simultaneously.

  2. ClinicalCodes: an online clinical codes repository to improve the validity and reproducibility of research using electronic medical records.

    PubMed

    Springate, David A; Kontopantelis, Evangelos; Ashcroft, Darren M; Olier, Ivan; Parisi, Rosa; Chamapiwa, Edmore; Reeves, David

    2014-01-01

    Lists of clinical codes are the foundation for research undertaken using electronic medical records (EMRs). If clinical code lists are not available, reviewers are unable to determine the validity of research, full study replication is impossible, researchers are unable to make effective comparisons between studies, and the construction of new code lists is subject to much duplication of effort. Despite this, the publication of clinical codes is rarely if ever a requirement for obtaining grants, validating protocols, or publishing research. In a representative sample of 450 EMR primary research articles indexed on PubMed, we found that only 19 (5.1%) were accompanied by a full set of published clinical codes and 32 (8.6%) stated that code lists were available on request. To help address these problems, we have built an online repository where researchers using EMRs can upload and download lists of clinical codes. The repository will enable clinical researchers to better validate EMR studies, build on previous code lists and compare disease definitions across studies. It will also assist health informaticians in replicating database studies, tracking changes in disease definitions or clinical coding practice through time and sharing clinical code information across platforms and data sources as research objects.

  3. ClinicalCodes: An Online Clinical Codes Repository to Improve the Validity and Reproducibility of Research Using Electronic Medical Records

    PubMed Central

    Springate, David A.; Kontopantelis, Evangelos; Ashcroft, Darren M.; Olier, Ivan; Parisi, Rosa; Chamapiwa, Edmore; Reeves, David

    2014-01-01

    Lists of clinical codes are the foundation for research undertaken using electronic medical records (EMRs). If clinical code lists are not available, reviewers are unable to determine the validity of research, full study replication is impossible, researchers are unable to make effective comparisons between studies, and the construction of new code lists is subject to much duplication of effort. Despite this, the publication of clinical codes is rarely if ever a requirement for obtaining grants, validating protocols, or publishing research. In a representative sample of 450 EMR primary research articles indexed on PubMed, we found that only 19 (5.1%) were accompanied by a full set of published clinical codes and 32 (8.6%) stated that code lists were available on request. To help address these problems, we have built an online repository where researchers using EMRs can upload and download lists of clinical codes. The repository will enable clinical researchers to better validate EMR studies, build on previous code lists and compare disease definitions across studies. It will also assist health informaticians in replicating database studies, tracking changes in disease definitions or clinical coding practice through time and sharing clinical code information across platforms and data sources as research objects. PMID:24941260

  4. Multi-level meta-workflows: new concept for regularly occurring tasks in quantum chemistry.

    PubMed

    Arshad, Junaid; Hoffmann, Alexander; Gesing, Sandra; Grunzke, Richard; Krüger, Jens; Kiss, Tamas; Herres-Pawlis, Sonja; Terstyanszky, Gabor

    2016-01-01

    In Quantum Chemistry, many tasks are reoccurring frequently, e.g. geometry optimizations, benchmarking series etc. Here, workflows can help to reduce the time of manual job definition and output extraction. These workflows are executed on computing infrastructures and may require large computing and data resources. Scientific workflows hide these infrastructures and the resources needed to run them. It requires significant efforts and specific expertise to design, implement and test these workflows. Many of these workflows are complex and monolithic entities that can be used for particular scientific experiments. Hence, their modification is not straightforward and it makes almost impossible to share them. To address these issues we propose developing atomic workflows and embedding them in meta-workflows. Atomic workflows deliver a well-defined research domain specific function. Publishing workflows in repositories enables workflow sharing inside and/or among scientific communities. We formally specify atomic and meta-workflows in order to define data structures to be used in repositories for uploading and sharing them. Additionally, we present a formal description focused at orchestration of atomic workflows into meta-workflows. We investigated the operations that represent basic functionalities in Quantum Chemistry, developed the relevant atomic workflows and combined them into meta-workflows. Having these workflows we defined the structure of the Quantum Chemistry workflow library and uploaded these workflows in the SHIWA Workflow Repository.Graphical AbstractMeta-workflows and embedded workflows in the template representation.

  5. Revision history aware repositories of computational models of biological systems.

    PubMed

    Miller, Andrew K; Yu, Tommy; Britten, Randall; Cooling, Mike T; Lawson, James; Cowan, Dougal; Garny, Alan; Halstead, Matt D B; Hunter, Peter J; Nickerson, David P; Nunns, Geo; Wimalaratne, Sarala M; Nielsen, Poul M F

    2011-01-14

    Building repositories of computational models of biological systems ensures that published models are available for both education and further research, and can provide a source of smaller, previously verified models to integrate into a larger model. One problem with earlier repositories has been the limitations in facilities to record the revision history of models. Often, these facilities are limited to a linear series of versions which were deposited in the repository. This is problematic for several reasons. Firstly, there are many instances in the history of biological systems modelling where an 'ancestral' model is modified by different groups to create many different models. With a linear series of versions, if the changes made to one model are merged into another model, the merge appears as a single item in the history. This hides useful revision history information, and also makes further merges much more difficult, as there is no record of which changes have or have not already been merged. In addition, a long series of individual changes made outside of the repository are also all merged into a single revision when they are put back into the repository, making it difficult to separate out individual changes. Furthermore, many earlier repositories only retain the revision history of individual files, rather than of a group of files. This is an important limitation to overcome, because some types of models, such as CellML 1.1 models, can be developed as a collection of modules, each in a separate file. The need for revision history is widely recognised for computer software, and a lot of work has gone into developing version control systems and distributed version control systems (DVCSs) for tracking the revision history. However, to date, there has been no published research on how DVCSs can be applied to repositories of computational models of biological systems. We have extended the Physiome Model Repository software to be fully revision history aware, by building it on top of Mercurial, an existing DVCS. We have demonstrated the utility of this approach, when used in conjunction with the model composition facilities in CellML, to build and understand more complex models. We have also demonstrated the ability of the repository software to present version history to casual users over the web, and to highlight specific versions which are likely to be useful to users. Providing facilities for maintaining and using revision history information is an important part of building a useful repository of computational models, as this information is useful both for understanding the source of and justification for parts of a model, and to facilitate automated processes such as merges. The availability of fully revision history aware repositories, and associated tools, will therefore be of significant benefit to the community.

  6. Applying knowledge-anchored hypothesis discovery methods to advance clinical and translational research: the OAMiner project

    PubMed Central

    Jackson, Rebecca D; Best, Thomas M; Borlawsky, Tara B; Lai, Albert M; James, Stephen; Gurcan, Metin N

    2012-01-01

    The conduct of clinical and translational research regularly involves the use of a variety of heterogeneous and large-scale data resources. Scalable methods for the integrative analysis of such resources, particularly when attempting to leverage computable domain knowledge in order to generate actionable hypotheses in a high-throughput manner, remain an open area of research. In this report, we describe both a generalizable design pattern for such integrative knowledge-anchored hypothesis discovery operations and our experience in applying that design pattern in the experimental context of a set of driving research questions related to the publicly available Osteoarthritis Initiative data repository. We believe that this ‘test bed’ project and the lessons learned during its execution are both generalizable and representative of common clinical and translational research paradigms. PMID:22647689

  7. Good Data Can Be Better Data - How Data Management Maturity Can Help Repositories Improve Operations, Data Quality, And Usability, Helping Researchers

    NASA Astrophysics Data System (ADS)

    Stall, S.

    2015-12-01

    Much earth and space science data and metadata are managed and supported by an infrastructure of repositories, ranging from large agency or instrument facilities, to institutions, to smaller repositories including labs. Scientists face many challenges in this ecosystem both on storing their data and in accessing data from others for new research. Critical for all uses is ensuring the credibility and integrity of the data and conveying that and provenance information now and in the future. Accurate information is essential for future researchers to find (or discover) the data, evaluate the data for use (content, temporal, geolocation, precision) and finally select (or discard) that data as meeting a "fit-for-purpose" criteria. We also need to optimize the effort it takes in describing the data for these determinations, which means making it efficient for the researchers who collect the data. At AGU we are developing a program aimed at helping repositories, and thereby researchers, improve data quality and data usability toward these goals. AGU has partnered with the CMMI Institute to develop their Data Management Maturity (DMM) framework within the Earth and space sciences. The CMMI DMM framework guides best practices in a range of data operations, and the application of the DMM, through an assessment, reveals how repositories and institutions can best optimize efforts to improve operations and functionality throughout the data lifecycle and elevate best practices across a variety of data management operations. Supporting processes like data operations, data governance, and data architecture are included. An assessment involves identifying accomplishment, and weaknesses compared to leading practices for data management. Broad application of the DMM can help improve quality in data and operations, and consistency across the community that will facilitate interoperability, discovery, preservation, and reuse. Good data can be better data. Consistency results in sustainability.

  8. Organizing the present, looking to the future: an online knowledge repository to facilitate collaboration.

    PubMed

    Burchill, C; Roos, L L; Fergusson, P; Jebamani, L; Turner, K; Dueck, S

    2000-01-01

    Comprehensive data available in the Canadian province of Manitoba since 1970 have aided study of the interaction between population health, health care utilization, and structural features of the health care system. Given a complex linked database and many ongoing projects, better organization of available epidemiological, institutional, and technical information was needed. The Manitoba Centre for Health Policy and Evaluation wished to develop a knowledge repository to handle data, document research Methods, and facilitate both internal communication and collaboration with other sites. This evolving knowledge repository consists of both public and internal (restricted access) pages on the World Wide Web (WWW). Information can be accessed using an indexed logical format or queried to allow entry at user-defined points. The main topics are: Concept Dictionary, Research Definitions, Meta-Index, and Glossary. The Concept Dictionary operationalizes concepts used in health research using administrative data, outlining the creation of complex variables. Research Definitions specify the codes for common surgical procedures, tests, and diagnoses. The Meta-Index organizes concepts and definitions according to the Medical Sub-Heading (MeSH) system developed by the National Library of Medicine. The Glossary facilitates navigation through the research terms and abbreviations in the knowledge repository. An Education Resources heading presents a web-based graduate course using substantial amounts of material in the Concept Dictionary, a lecture in the Epidemiology Supercourse, and material for Manitoba's Regional Health Authorities. Confidential information (including Data Dictionaries) is available on the Centre's internal website. Use of the public pages has increased dramatically since January 1998, with almost 6,000 page hits from 250 different hosts in May 1999. More recently, the number of page hits has averaged around 4,000 per month, while the number of unique hosts has climbed to around 400. This knowledge repository promotes standardization and increases efficiency by placing concepts and associated programming in the Centre's collective memory. Collaboration and project management are facilitated.

  9. Organizing the Present, Looking to the Future: An Online Knowledge Repository to Facilitate Collaboration

    PubMed Central

    Burchill, Charles; Fergusson, Patricia; Jebamani, Laurel; Turner, Ken; Dueck, Stephen

    2000-01-01

    Background Comprehensive data available in the Canadian province of Manitoba since 1970 have aided study of the interaction between population health, health care utilization, and structural features of the health care system. Given a complex linked database and many ongoing projects, better organization of available epidemiological, institutional, and technical information was needed. Objective The Manitoba Centre for Health Policy and Evaluation wished to develop a knowledge repository to handle data, document research methods, and facilitate both internal communication and collaboration with other sites. Methods This evolving knowledge repository consists of both public and internal (restricted access) pages on the World Wide Web (WWW). Information can be accessed using an indexed logical format or queried to allow entry at user-defined points. The main topics are: Concept Dictionary, Research Definitions, Meta-Index, and Glossary. The Concept Dictionary operationalizes concepts used in health research using administrative data, outlining the creation of complex variables. Research Definitions specify the codes for common surgical procedures, tests, and diagnoses. The Meta-Index organizes concepts and definitions according to the Medical Sub-Heading (MeSH) system developed by the National Library of Medicine. The Glossary facilitates navigation through the research terms and abbreviations in the knowledge repository. An Education Resources heading presents a web-based graduate course using substantial amounts of material in the Concept Dictionary, a lecture in the Epidemiology Supercourse, and material for Manitoba's Regional Health Authorities. Confidential information (including Data Dictionaries) is available on the Centre's internal website. Results Use of the public pages has increased dramatically since January 1998, with almost 6,000 page hits from 250 different hosts in May 1999. More recently, the number of page hits has averaged around 4,000 per month, while the number of unique hosts has climbed to around 400. Conclusions This knowledge repository promotes standardization and increases efficiency by placing concepts and associated programming in the Centre's collective memory. Collaboration and project management are facilitated. PMID:11720929

  10. PGP repository: a plant phenomics and genomics data publication infrastructure.

    PubMed

    Arend, Daniel; Junker, Astrid; Scholz, Uwe; Schüler, Danuta; Wylie, Juliane; Lange, Matthias

    2016-01-01

    Plant genomics and phenomics represents the most promising tools for accelerating yield gains and overcoming emerging crop productivity bottlenecks. However, accessing this wealth of plant diversity requires the characterization of this material using state-of-the-art genomic, phenomic and molecular technologies and the release of subsequent research data via a long-term stable, open-access portal. Although several international consortia and public resource centres offer services for plant research data management, valuable digital assets remains unpublished and thus inaccessible to the scientific community. Recently, the Leibniz Institute of Plant Genetics and Crop Plant Research and the German Plant Phenotyping Network have jointly initiated the Plant Genomics and Phenomics Research Data Repository (PGP) as infrastructure to comprehensively publish plant research data. This covers in particular cross-domain datasets that are not being published in central repositories because of its volume or unsupported data scope, like image collections from plant phenotyping and microscopy, unfinished genomes, genotyping data, visualizations of morphological plant models, data from mass spectrometry as well as software and documents.The repository is hosted at Leibniz Institute of Plant Genetics and Crop Plant Research using e!DAL as software infrastructure and a Hierarchical Storage Management System as data archival backend. A novel developed data submission tool was made available for the consortium that features a high level of automation to lower the barriers of data publication. After an internal review process, data are published as citable digital object identifiers and a core set of technical metadata is registered at DataCite. The used e!DAL-embedded Web frontend generates for each dataset a landing page and supports an interactive exploration. PGP is registered as research data repository at BioSharing.org, re3data.org and OpenAIRE as valid EU Horizon 2020 open data archive. Above features, the programmatic interface and the support of standard metadata formats, enable PGP to fulfil the FAIR data principles-findable, accessible, interoperable, reusable.Database URL:http://edal.ipk-gatersleben.de/repos/pgp/. © The Author(s) 2016. Published by Oxford University Press.

  11. Cognitive Challenges

    MedlinePlus

    ... Find Local Resources Publications Webinars and Videos Biosample Repository Patient-Focused Drug Development Learn Engage Donate Healthcare ... and Funding Preclinical Research Natural History Database Biosample ... Research Consortium Research Conferences Research Resources International ...

  12. Training and Best Practice Guidelines: Implications for Metadata Creation

    ERIC Educational Resources Information Center

    Chuttur, Mohammad Y.

    2012-01-01

    In response to the rapid development of digital libraries over the past decade, researchers have focused on the use of metadata as an effective means to support resource discovery within online repositories. With the increasing involvement of libraries in digitization projects and the growing number of institutional repositories, it is anticipated…

  13. Content Analysis of a Computer-Based Faculty Activity Repository

    ERIC Educational Resources Information Center

    Baker-Eveleth, Lori; Stone, Robert W.

    2013-01-01

    The research presents an analysis of faculty opinions regarding the introduction of a new computer-based faculty activity repository (FAR) in a university setting. The qualitative study employs content analysis to better understand the phenomenon underlying these faculty opinions and to augment the findings from a quantitative study. A web-based…

  14. XNAT Central: Open sourcing imaging research data.

    PubMed

    Herrick, Rick; Horton, William; Olsen, Timothy; McKay, Michael; Archie, Kevin A; Marcus, Daniel S

    2016-01-01

    XNAT Central is a publicly accessible medical imaging data repository based on the XNAT open-source imaging informatics platform. It hosts a wide variety of research imaging data sets. The primary motivation for creating XNAT Central was to provide a central repository to host and provide access to a wide variety of neuroimaging data. In this capacity, XNAT Central hosts a number of data sets from research labs and investigative efforts from around the world, including the OASIS Brains imaging studies, the NUSDAST study of schizophrenia, and more. Over time, XNAT Central has expanded to include imaging data from many different fields of research, including oncology, orthopedics, cardiology, and animal studies, but continues to emphasize neuroimaging data. Through the use of XNAT's DICOM metadata extraction capabilities, XNAT Central provides a searchable repository of imaging data that can be referenced by groups, labs, or individuals working in many different areas of research. The future development of XNAT Central will be geared towards greater ease of use as a reference library of heterogeneous neuroimaging data and associated synthetic data. It will also become a tool for making data available supporting published research and academic articles. Copyright © 2015 Elsevier Inc. All rights reserved.

  15. Digital Repositories and the Question of Data Usefulness

    NASA Astrophysics Data System (ADS)

    Hughes, J. S.; Downs, R. R.

    2017-12-01

    The advent of ISO standards for trustworthy long-term digital repositories provides both a set of principles to develop long-term data repositories and the instruments to assess them for trustworthiness. Such mandatory high-level requirements are broad enough to be achievable, to some extent, by many scientific data centers, archives, and other repositories. But the requirement that the data be useful in the future, the requirement that is usually considered to be most relevant to the value of the repository for its user communities, largely remains subject to various interpretations and misunderstanding. However, current and future users will be relying on repositories to preserve and disseminate the data and information needed to discover, understand, and utilize these resources to support their research, learning, and decision-making objectives. Therefore, further study is needed to determine the approaches that can be adopted by repositories to make data useful to future communities of users. This presentation will describe approaches for enabling scientific data and related information, such as software, to be useful for current and potential future user communities and will present the methodology chosen to make one science discipline's data useful for both current and future users. The method uses an ontology-based information model to define and capture the information necessary to make the data useful for contemporary and future users.

  16. Lungs in TSC

    MedlinePlus

    ... Find Local Resources Publications Webinars and Videos Biosample Repository Patient-Focused Drug Development Learn Engage Donate Healthcare ... and Funding Preclinical Research Natural History Database Biosample ... Research Consortium Research Conferences Research Resources International ...

  17. 10 CFR 60.44 - Changes, tests, and experiments.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 10 Energy 2 2011-01-01 2011-01-01 false Changes, tests, and experiments. 60.44 Section 60.44... REPOSITORIES Licenses License Issuance and Amendment § 60.44 Changes, tests, and experiments. (a)(1) Following... experiments not described in the application, without prior Commission approval, provided the change, test, or...

  18. 10 CFR 60.44 - Changes, tests, and experiments.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 10 Energy 2 2012-01-01 2012-01-01 false Changes, tests, and experiments. 60.44 Section 60.44... REPOSITORIES Licenses License Issuance and Amendment § 60.44 Changes, tests, and experiments. (a)(1) Following... experiments not described in the application, without prior Commission approval, provided the change, test, or...

  19. 10 CFR 60.44 - Changes, tests, and experiments.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 10 Energy 2 2013-01-01 2013-01-01 false Changes, tests, and experiments. 60.44 Section 60.44... REPOSITORIES Licenses License Issuance and Amendment § 60.44 Changes, tests, and experiments. (a)(1) Following... experiments not described in the application, without prior Commission approval, provided the change, test, or...

  20. 10 CFR 60.44 - Changes, tests, and experiments.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 10 Energy 2 2014-01-01 2014-01-01 false Changes, tests, and experiments. 60.44 Section 60.44... REPOSITORIES Licenses License Issuance and Amendment § 60.44 Changes, tests, and experiments. (a)(1) Following... experiments not described in the application, without prior Commission approval, provided the change, test, or...

  1. Workshop on development of radionuclide getters for the Yucca Mountain waste repository: proceedings.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Moore, Robert Charles; Lukens, Wayne W.

    The proposed Yucca Mountain repository, located in southern Nevada, is to be the first facility for permanent disposal of spent reactor fuel and high-level radioactive waste in the United States. Total Systems Performance Assessment (TSPA) analysis has indicated that among the major radionuclides contributing to dose are technetium, iodine, and neptunium, all of which are highly mobile in the environment. Containment of these radionuclides within the repository is a priority for the Yucca Mountain Project (YMP). These proceedings review current research and technology efforts for sequestration of the radionuclides with a focus on technetium, iodine, and neptunium. This workshop alsomore » covered issues concerning the Yucca Mountain environment and getter characteristics required for potential placement into the repository.« less

  2. Repository Profiles for Atmospheric and Climate Sciences: Capabilities and Trends in Data Services

    NASA Astrophysics Data System (ADS)

    Hou, C. Y.; Thompson, C. A.; Palmer, C. L.

    2014-12-01

    As digital research data proliferate and expectations for open access escalate, the landscape of data repositories is becoming more complex. For example, DataBib currently identifies 980 data repositories across the disciplines, with 117 categorized under Geosciences. In atmospheric and climate sciences, there are great expectations for the integration and reuse of data for advancing science. To realize this potential, resources are needed that explicate the range of repository options available for locating and depositing open data, their conditions of access and use, and the services and tools they provide. This study profiled 38 open digital repositories in the atmospheric and climate sciences, analyzing each on 55 criteria through content analysis of their websites. The results provide a systematic way to assess and compare capabilities, services, and institutional characteristics and identify trends across repositories. Selected results from the more detailed outcomes to be presented: Most repositories offer guidance on data format(s) for submission and dissemination. 42% offer authorization-free access. More than half use some type of data identification system such as DOIs. Nearly half offer some data processing, with a similar number providing software or tools. 78.9% request that users cite or acknowledge datasets used and the data center. Only 21.1% recommend specific metadata standards, such as ISO 19115 or Dublin Core, with more than half utilizing a customized metadata scheme. Information was rarely provided on repository certification and accreditation and uneven for transfer of rights and data security. Few provided policy information on preservation, migration, reappraisal, disposal, or long-term sustainability. As repository use increases, it will be important for institutions to make their procedures and policies explicit, to build trust with user communities and improve efficiencies in data sharing. Resources such as repository profiles will be essential for scientists to weigh options and understand trends in data services across the evolving network of repositories.

  3. Eye Involvement in TSC

    MedlinePlus

    ... Find Local Resources Publications Webinars and Videos Biosample Repository Patient-Focused Drug Development Learn Engage Donate Healthcare ... and Funding Preclinical Research Natural History Database Biosample ... Research Consortium Research Conferences Research Resources International ...

  4. NUTRITION w/Repository

    NASA Image and Video Library

    2009-06-06

    ISS020-E-007603 (7 June 2009) --- European Space Agency astronaut Frank De Winne, Expedition 20 flight engineer, removes a dewar tray from the Minus Eighty Laboratory Freezer for ISS (MELFI) in order to insert biological samples into the trays in the Kibo laboratory of the International Space Station. Samples were taken as part of the Nutritional Status Assessment (Nutrition) with Repository experiment, a study done by NASA to date of human physiologic changes during long-duration spaceflight.

  5. Using neural networks in software repositories

    NASA Technical Reports Server (NTRS)

    Eichmann, David (Editor); Srinivas, Kankanahalli; Boetticher, G.

    1992-01-01

    The first topic is an exploration of the use of neural network techniques to improve the effectiveness of retrieval in software repositories. The second topic relates to a series of experiments conducted to evaluate the feasibility of using adaptive neural networks as a means of deriving (or more specifically, learning) measures on software. Taken together, these two efforts illuminate a very promising mechanism supporting software infrastructures - one based upon a flexible and responsive technology.

  6. The Nevada initiative: A risk communication Fiasco

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Flynn, J.; Solvic, P.; Mertz, C.K.

    The U.S. Congress has designated Yucca Mountain, Nevada as the only potential site to be studied for the nation`s first high-level nuclear waste repository. People in Nevada strongly oppose the program, managed by the U.S. Department of Energy. Survey research shows that the public believes there are great risks from a repository program, in contrast to a majority of scientists who feel the risks are acceptably small. Delays in the repository program resulting in part from public opposition in Nevada have concerned the nuclear power industry, which collects the fees for the federal repository program and believes it needs themore » repository as a final disposal facility for its high-level nuclear wastes. To assist the repository program, the American Nuclear Energy Council (ANEC), an industry group, sponsored a massive advertising campaign in Nevada. The campaign attempted to assure people that the risks of a repository were small and that the repository studies should proceed. The campaign failed because its managers misunderstood the issues underlying the controversy, attempted a covert manipulation of public opinion that was revealed, and most importantly, lacked the public trust that was necessary to communicate credibly about the risks of a nuclear waste facility. This article describes the advertising campaign and its effects. The manner in which the ANEC campaign itself became a controversial public issue is reviewed. The advertising campaign is discussed as it relates to risk assessment and communication. 29 refs., 2 tabs.« less

  7. How Elsevier is supporting the value and usefulness of data with Cross-linking and Research Data Services.

    NASA Astrophysics Data System (ADS)

    Keall, Bethan; Koers, Hylke; Marques, David

    2013-04-01

    Research in the Earth & Planetary Sciences is characterized by a wealth of observational data - ranging from observations by satellites orbiting the Earth, to borehole measurements at the bottom of the ocean, and also includes data from projects like the Rover Curiosity Landing. Thanks to technological advancements, it has become much easier for researchers over the last few decades to gather large volumes of data, analyze, and share with other researchers inside and outside the lab. With data serving such an important role in the way research is carried out, it becomes a crucial task to archive, maintain, organize, and disseminate research data in a dependable and structured manner. Subject-specific data repositories, often driven by the scientific community, are taking an increasingly prominent role in this domain, getting traction amongst researchers as the go-to place to deposit raw research data. At the same time, the scientific article remains an essential resource of scientific information. At Elsevier, we strive to continuously adapt the article format to meet the needs of modern-day researchers. This includes better support for digital content (see, e.g., http://www.elsevier.com/googlemaps), but also bidirectional linking between online articles and data repositories. In this spirit, Elsevier is collaborating with several leading data repositories, such as PANGAEA, IEDA, and NERC, to interlink articles and data for improved visibility and discoverability of both primary research data and research articles. In addition, Elsevier has formed a new group, Research Data Services, with three primary goals: • help increase the sharing and archiving of research data in discipline-specific repositories • help increase the value of shared data, particularly with annotation and provenance metadata and linking discipline-specific datasets together • help create a credit and impact assessment infrastructure to make research data independently important in its own right. We are working on several initiatives at Elsevier that enhance the online article format, and to make it easier for researchers to share, find, access, link together and analyze relevant research data. This helps to increase the value of both articles and data, and enables researchers to gain full credit for their research data output.

  8. Earth Observation Data Quality Monitoring and Control: A Case Study of STAR Central Data Repository

    NASA Astrophysics Data System (ADS)

    Han, W.; Jochum, M.

    2017-12-01

    Earth observation data quality is very important for researchers and decision makers involved in weather forecasting, severe weather warning, disaster and emergency response, environmental monitoring, etc. Monitoring and control earth observation data quality, especially accuracy, completeness, and timeliness, is very useful in data management and governance to optimize data flow, discover potential transmission issues, and better connect data providers and users. Taking a centralized near real-time satellite data repository, STAR (Center for Satellite Applications and Research of NOAA) Central Data Repository (SCDR), as an example, this paper describes how to develop new mechanism to verify data integrity, check data completeness, and monitor data latency in an operational data management system. Such quality monitoring and control of large volume satellite data help data providers and managers improve data transmission of near real-time satellite data, enhance its acquisition and management, and overcome performance and management issues to better serve research and development activities.

  9. The United States Polar Rock Repository: A geological resource for the Earth science community

    USGS Publications Warehouse

    Grunow, Annie M.; Elliot, David H.; Codispoti, Julie E.

    2007-01-01

    The United States Polar Rock Repository (USPRR) is a U. S. national facility designed for the permanent curatorial preservation of rock samples, along with associated materials such as field notes, annotated air photos and maps, raw analytic data, paleomagnetic cores, ground rock and mineral residues, thin sections, and microfossil mounts, microslides and residues from Polar areas. This facility was established by the Office of Polar Programs at the U. S. National Science Foundation (NSF) to minimize redundant sample collecting, and also because the extreme cold and hazardous field conditions make fieldwork costly and difficult. The repository provides, along with an on-line database of sample information, an essential resource for proposal preparation, pilot studies and other sample based research that should make fieldwork more efficient and effective. This latter aspect should reduce the environmental impact of conducting research in sensitive Polar Regions. The USPRR also provides samples for educational outreach. Rock samples may be borrowed for research or educational purposes as well as for museum exhibits.

  10. Central Satellite Data Repository Supporting Research and Development

    NASA Astrophysics Data System (ADS)

    Han, W.; Brust, J.

    2015-12-01

    Near real-time satellite data is critical to many research and development activities of atmosphere, land, and ocean processes. Acquiring and managing huge volumes of satellite data without (or with less) latency in an organization is always a challenge in the big data age. An organization level data repository is a practical solution to meeting this challenge. The STAR (Center for Satellite Applications and Research of NOAA) Central Data Repository (SCDR) is a scalable, stable, and reliable repository to acquire, manipulate, and disseminate various types of satellite data in an effective and efficient manner. SCDR collects more than 200 data products, which are commonly used by multiple groups in STAR, from NOAA, GOES, Metop, Suomi NPP, Sentinel, Himawari, and other satellites. The processes of acquisition, recording, retrieval, organization, and dissemination are performed in parallel. Multiple data access interfaces, like FTP, FTPS, HTTP, HTTPS, and RESTful, are supported in the SCDR to obtain satellite data from their providers through high speed internet. The original satellite data in various raster formats can be parsed in the respective adapter to retrieve data information. The data information is ingested to the corresponding partitioned tables in the central database. All files are distributed equally on the Network File System (NFS) disks to balance the disk load. SCDR provides consistent interfaces (including Perl utility, portal, and RESTful Web service) to locate files of interest easily and quickly and access them directly by over 200 compute servers via NFS. SCDR greatly improves collection and integration of near real-time satellite data, addresses satellite data requirements of scientists and researchers, and facilitates their primary research and development activities.

  11. Exploring a New Model for Preprint Server: A Case Study of CSPO

    ERIC Educational Resources Information Center

    Hu, Changping; Zhang, Yaokun; Chen, Guo

    2010-01-01

    This paper describes the introduction of an open-access preprint server in China covering 43 disciplines. The system includes mandatory deposit for state-funded research and reports on the repository and its effectiveness and outlines a novel process of peer-review of preprints in the repository, which can be incorporated into the established…

  12. Social Influences on User Behavior in Group Information Repositories

    ERIC Educational Resources Information Center

    Rader, Emilee Jeanne

    2009-01-01

    Group information repositories are systems for organizing and sharing files kept in a central location that all group members can access. These systems are often assumed to be tools for storage and control of files and their metadata, not tools for communication. The purpose of this research is to better understand user behavior in group…

  13. YUCCA MOUNTAIN: Earth-Science Issues at a Geologic Repository for High-Level Nuclear Waste

    NASA Astrophysics Data System (ADS)

    Long, Jane C. S.

    2004-05-01

    The nation has over 40,000 metric tonnes (MT) of nuclear waste destined for disposal in a geologic repository at Yucca Mountain. In this review, we highlight some of the important geoscience issues associated with the project and place them in the context of the process by which a final decision on Yucca Mountain will be made. The issues include understanding how water could infiltrate the repository, corrode the canisters, dissolve the waste, and transport it to the biosphere during a 10,000-year compliance period in a region, the Basin and Range province, that is known for seismic and volcanic activity. Although the site is considered to be "dry," a considerable amount of water is present as pore waters and as structural water in zeolites. The geochemical environment is oxidizing, and the present repository design will maintain temperatures at greater than 100°C for thousands of years. Geoscientists in this project are challenged to make unprecedented predictions about coupled thermal, hydrologic, mechanical, and geochemical processes governing the future behavior of the repository and to conduct research in a regulatory and legal environment that requires a quantitative analysis of repository performance.

  14. GeoTrust Hub: A Platform For Sharing And Reproducing Geoscience Applications

    NASA Astrophysics Data System (ADS)

    Malik, T.; Tarboton, D. G.; Goodall, J. L.; Choi, E.; Bhatt, A.; Peckham, S. D.; Foster, I.; Ton That, D. H.; Essawy, B.; Yuan, Z.; Dash, P. K.; Fils, G.; Gan, T.; Fadugba, O. I.; Saxena, A.; Valentic, T. A.

    2017-12-01

    Recent requirements of scholarly communication emphasize the reproducibility of scientific claims. Text-based research papers are considered poor mediums to establish reproducibility. Papers must be accompanied by "research objects", aggregation of digital artifacts that together with the paper provide an authoritative record of a piece of research. We will present GeoTrust Hub (http://geotrusthub.org), a platform for creating, sharing, and reproducing reusable research objects. GeoTrust Hub provides tools for scientists to create `geounits'--reusable research objects. Geounits are self-contained, annotated, and versioned containers that describe and package computational experiments in an efficient and light-weight manner. Geounits can be shared on public repositories such as HydroShare and FigShare, and also using their respective APIs reproduced on provisioned clouds. The latter feature enables science applications to have a lifetime beyond sharing, wherein they can be independently verified and trust be established as they are repeatedly reused. Through research use cases from several geoscience laboratories across the United States, we will demonstrate how tools provided from GeoTrust Hub along with Hydroshare as its public repository for geounits is advancing the state of reproducible research in the geosciences. For each use case, we will address different computational reproducibility requirements. Our first use case will be an example of setup reproducibility which enables a scientist to set up and reproduce an output from a model with complex configuration and development environments. Our second use case will be an example of algorithm/data reproducibility, where in a shared data science model/dataset can be substituted with an alternate one to verify model output results, and finally an example of interactive reproducibility, in which an experiment is dependent on specific versions of data to produce the result. Toward this we will use software and data used in preparing data for the MODFLOW model in Hydrology, JupyterHub used in Hydroshare, PyLith used in Computational Infrastructure for Geodynamics, and GeoSpace Collaborative Observations and Assimilative Modeling used in space science. The GeoTrust Hub is funded through the National Science Foundation EarthCube program.

  15. The Galileo Teacher Training Programme

    NASA Astrophysics Data System (ADS)

    Doran, Rosa

    The Galileo Teacher Training Program is a global effort to empower teachers all over the world to embark on a new trend in science teaching, using new technologies and real research meth-ods to teach curriculum content. The GTTP goal is to create a worldwide network of "Galileo Ambassadors", promoters of GTTP training session, and a legion of "Galileo Teachers", edu-cators engaged on the use of innovative resources and sharing experiences and supporting its pears worldwide. Through workshops, online training tools and resources, the products and techniques promoted by this program can be adapted to reach locations with few resources of their own, as well as network-connected areas that can take advantage of access to robotic, optical and radio telescopes, webcams, astronomy exercises, cross-disciplinary resources, image processing and digital universes (web and desktop planetariums). Promoters of GTTP are expert astronomy educators connected to Universities or EPO institutions that facilitate the consolidation of an active support to newcomers and act as a 24 hour helpdesk to teachers all over the world. GTTP will also engage in the creation of a repository of astronomy education resources and science research projects, ViRoS (Virtual Repository of resources and Science Projects), in order to simplify the task of educators willing to enrich classroom activities.

  16. BioSurfDB: knowledge and algorithms to support biosurfactants and biodegradation studies

    PubMed Central

    Oliveira, Jorge S.; Araújo, Wydemberg; Lopes Sales, Ana Isabela; de Brito Guerra, Alaine; da Silva Araújo, Sinara Carla; de Vasconcelos, Ana Tereza Ribeiro; Agnez-Lima, Lucymara F.; Freitas, Ana Teresa

    2015-01-01

    Crude oil extraction, transportation and use provoke the contamination of countless ecosystems. Therefore, bioremediation through surfactants mobilization or biodegradation is an important subject, both economically and environmentally. Bioremediation research had a great boost with the recent advances in Metagenomics, as it enabled the sequencing of uncultured microorganisms providing new insights on surfactant-producing and/or oil-degrading bacteria. Many research studies are making available genomic data from unknown organisms obtained from metagenomics analysis of oil-contaminated environmental samples. These new datasets are presently demanding the development of new tools and data repositories tailored for the biological analysis in a context of bioremediation data analysis. This work presents BioSurfDB, www.biosurfdb.org, a curated relational information system integrating data from: (i) metagenomes; (ii) organisms; (iii) biodegradation relevant genes; proteins and their metabolic pathways; (iv) bioremediation experiments results, with specific pollutants treatment efficiencies by surfactant producing organisms; and (v) a biosurfactant-curated list, grouped by producing organism, surfactant name, class and reference. The main goal of this repository is to gather information on the characterization of biological compounds and mechanisms involved in biosurfactant production and/or biodegradation and make it available in a curated way and associated with a number of computational tools to support studies of genomic and metagenomic data. Database URL: www.biosurfdb.org PMID:25833955

  17. Coupled Biological-Geomechanical-Geochemical Effects of the Disturbed Rock Zone on the Performance of the Waste Isolation Pilot Plant

    NASA Astrophysics Data System (ADS)

    Dunagan, S. C.; Herrick, C. G.; Lee, M. Y.

    2008-12-01

    The Waste Isolation Pilot Plant (WIPP) is located at a depth of 655 m in bedded salt in southeastern New Mexico and is operated by the U.S. Department of Energy as a deep underground disposal facility for transuranic (TRU) waste. The WIPP must comply with the EPA's environmental regulations that require a probabilistic risk analysis of releases of radionuclides due to inadvertent human intrusion into the repository at some time during the 10,000-year regulatory period. Sandia National Laboratories conducts performance assessments (PAs) of the WIPP using a system of computer codes representing the evolution of underground repository and emplaced TRU waste in order to demonstrate compliance. One of the important features modeled in a PA is the disturbed rock zone (DRZ) surrounding the emplacement rooms in the repository. The extent and permeability of DRZ play a significant role in the potential radionuclide release scenarios. We evaluated the phenomena occurring in the repository that affect the DRZ and their potential effects on the extent and permeability of the DRZ. Furthermore, we examined the DRZ's role in determining the performance of the repository. Pressure in the completely sealed repository will be increased by creep closure of the salt and degradation of TRU waste contents by microbial activity in the repository. An increased pressure in the repository will reduce the extent and permeability of the DRZ. The reduced DRZ extent and permeability will decrease the amount of brine that is available to interact with the waste. Furthermore, the potential for radionuclide release from the repository is dependent on the amount of brine that enters the repository. As a result of these coupled biological-geomechanical-geochemical phenomena, the extent and permeability of the DRZ has a significant impact on the potential radionuclide releases from the repository and, in turn, the repository performance. Sandia is a multi program laboratory operated by Sandia Corporation, a Lockheed Martin Company, for the United States Department of Energy's National Nuclear Security Administration under Contract DE-AC04- 94AL85000. This research is funded by WIPP programs administered by the Office of Environmental Management (EM) of the U.S. Department of Energy.

  18. Semantic Technologies and Bio-Ontologies.

    PubMed

    Gutierrez, Fernando

    2017-01-01

    As information available through data repositories constantly grows, the need for automated mechanisms for linking, querying, and sharing data has become a relevant factor both in research and industry. This situation is more evident in research fields such as the life sciences, where new experiments by different research groups are constantly generating new information regarding a wide variety of related study objects. However, current methods for representing information and knowledge are not suited for machine processing. The Semantic Technologies are a set of standards and protocols that intend to provide methods for representing and handling data that encourages reusability of information and is machine-readable. In this chapter, we will provide a brief introduction to Semantic Technologies, and how these protocols and standards have been incorporated into the life sciences to facilitate dissemination and access to information.

  19. Semantic Web repositories for genomics data using the eXframe platform

    PubMed Central

    2014-01-01

    Background With the advent of inexpensive assay technologies, there has been an unprecedented growth in genomics data as well as the number of databases in which it is stored. In these databases, sample annotation using ontologies and controlled vocabularies is becoming more common. However, the annotation is rarely available as Linked Data, in a machine-readable format, or for standardized queries using SPARQL. This makes large-scale reuse, or integration with other knowledge bases very difficult. Methods To address this challenge, we have developed the second generation of our eXframe platform, a reusable framework for creating online repositories of genomics experiments. This second generation model now publishes Semantic Web data. To accomplish this, we created an experiment model that covers provenance, citations, external links, assays, biomaterials used in the experiment, and the data collected during the process. The elements of our model are mapped to classes and properties from various established biomedical ontologies. Resource Description Framework (RDF) data is automatically produced using these mappings and indexed in an RDF store with a built-in Sparql Protocol and RDF Query Language (SPARQL) endpoint. Conclusions Using the open-source eXframe software, institutions and laboratories can create Semantic Web repositories of their experiments, integrate it with heterogeneous resources and make it interoperable with the vast Semantic Web of biomedical knowledge. PMID:25093072

  20. Container Approval for the Disposal of Radioactive Waste with Negligible Heat Generation in the German Konrad Repository - 12148

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Voelzke, Holger; Nieslony, Gregor; Ellouz, Manel

    Since the license for the Konrad repository was finally confirmed by legal decision in 2007, the Federal Institute for Radiation Protection (BfS) has been performing further planning and preparation work to prepare the repository for operation. Waste conditioning and packaging has been continued by different waste producers as the nuclear industry and federal research institutes on the basis of the official disposal requirements. The necessary prerequisites for this are approved containers as well as certified waste conditioning and packaging procedures. The Federal Institute for Materials Research and Testing (BAM) is responsible for container design testing and evaluation of quality assurancemore » measures on behalf of BfS under consideration of the Konrad disposal requirements. Besides assessing the container handling stability (stacking tests, handling loads), design testing procedures are performed that include fire tests (800 deg. C, 1 hour) and drop tests from different heights and drop orientations. This paper presents the current state of BAM design testing experiences about relevant container types (box shaped, cylindrical) made of steel sheets, ductile cast iron or concrete. It explains usual testing and evaluation methods which range from experimental testing to analytical and numerical calculations. Another focus has been laid on already existing containers and packages. The question arises as to how they can be evaluated properly especially with respect to lack of completeness of safety assessment and fabrication documentation. At present BAM works on numerous applications for container design testing for the Konrad repository. Some licensing procedures were successfully finished in the past and BfS certified several container types like steel sheet, concrete until cast iron containers which are now available for waste packaging for final disposal. However, large quantities of radioactive wastes had been placed into interim storage using containers which are not already licensed for the Konrad repository. Safety assessment of these so-called 'old' containers is a big challenge for all parties because documentation sheets about container design testing and fabrication often contain gaps or have not yet been completed. Appropriate solution strategies are currently under development and discussion. Furthermore, BAM has successfully initiated and established an information forum, called 'ERFA QM Konrad Containers', which facilitates discussions on various issues of common interest with respect to Konrad container licensing procedures as well as the interpretation of disposal requirements under consideration of operational needs. Thus, it provides additional, valuable supports for container licensing procedures. (authors)« less

  1. Types of Seizures Affecting Individuals with TSC

    MedlinePlus

    ... Find Local Resources Publications Webinars and Videos Biosample Repository Patient-Focused Drug Development Learn Engage Donate Healthcare ... and Funding Preclinical Research Natural History Database Biosample ... Research Consortium Research Conferences Research Resources International ...

  2. Information warehouse - a comprehensive informatics platform for business, clinical, and research applications.

    PubMed

    Kamal, Jyoti; Liu, Jianhua; Ostrander, Michael; Santangelo, Jennifer; Dyta, Ravi; Rogers, Patrick; Mekhjian, Hagop S

    2010-11-13

    Since its inception in 1997, the IW (Information Warehouse) at the Ohio State University Medical Center (OSUMC) has gradually transformed itself from a single purpose business decision support system to a comprehensive informatics platform supporting basic, clinical, and translational research. The IW today is the combination of four integrated components: a clinical data repository containing over a million patients; a research data repository housing various research specific data; an application development platform for building business and research enabling applications; a business intelligence environment assisting in reporting in all function areas. The IW is structured and encoded using standard terminologies such as SNOMED-CT, ICD, and CPT. The IW is an important component of OSUMC's Clinical and Translational Science Award (CTSA) informatics program.

  3. Digital Rocks Portal: a Sustainable Platform for Data Management, Analysis and Remote Visualization of Volumetric Images of Porous Media

    NASA Astrophysics Data System (ADS)

    Prodanovic, M.; Esteva, M.; Ketcham, R. A.

    2017-12-01

    Nanometer to centimeter-scale imaging such as (focused ion beam) scattered electron microscopy, magnetic resonance imaging and X-ray (micro)tomography has since 1990s introduced 2D and 3D datasets of rock microstructure that allow investigation of nonlinear flow and mechanical phenomena on the length scales that are otherwise impervious to laboratory measurements. The numerical approaches that use such images produce various upscaled parameters required by subsurface flow and deformation simulators. All of this has revolutionized our knowledge about grain scale phenomena. However, a lack of data-sharing infrastructure among research groups makes it difficult to integrate different length scales. We have developed a sustainable, open and easy-to-use repository called the Digital Rocks Portal (https://www.digitalrocksportal.org), that (1) organizes images and related experimental measurements of different porous materials, (2) improves access to them for a wider community of engineering or geosciences researchers not necessarily trained in computer science or data analysis. Digital Rocks Portal (NSF EarthCube Grant 1541008) is the first repository for imaged porous microstructure data. It is implemented within the reliable, 24/7 maintained High Performance Computing Infrastructure supported by the Texas Advanced Computing Center (University of Texas at Austin). Long-term storage is provided through the University of Texas System Research Cyber-infrastructure initiative. We show how the data can be documented, referenced in publications via digital object identifiers (see Figure below for examples), visualized, searched for and linked to other repositories. We show recently implemented integration of the remote parallel visualization, bulk upload for large datasets as well as preliminary flow simulation workflow with the pore structures currently stored in the repository. We discuss the issues of collecting correct metadata, data discoverability and repository sustainability.

  4. Construction of a nasopharyngeal carcinoma 2D/MS repository with Open Source XML database--Xindice.

    PubMed

    Li, Feng; Li, Maoyu; Xiao, Zhiqiang; Zhang, Pengfei; Li, Jianling; Chen, Zhuchu

    2006-01-11

    Many proteomics initiatives require integration of all information with uniformcriteria from collection of samples and data display to publication of experimental results. The integration and exchanging of these data of different formats and structure imposes a great challenge to us. The XML technology presents a promise in handling this task due to its simplicity and flexibility. Nasopharyngeal carcinoma (NPC) is one of the most common cancers in southern China and Southeast Asia, which has marked geographic and racial differences in incidence. Although there are some cancer proteome databases now, there is still no NPC proteome database. The raw NPC proteome experiment data were captured into one XML document with Human Proteome Markup Language (HUP-ML) editor and imported into native XML database Xindice. The 2D/MS repository of NPC proteome was constructed with Apache, PHP and Xindice to provide access to the database via Internet. On our website, two methods, keyword query and click query, were provided at the same time to access the entries of the NPC proteome database. Our 2D/MS repository can be used to share the raw NPC proteomics data that are generated from gel-based proteomics experiments. The database, as well as the PHP source codes for constructing users' own proteome repository, can be accessed at http://www.xyproteomics.org/.

  5. Metadata management and semantics in microarray repositories.

    PubMed

    Kocabaş, F; Can, T; Baykal, N

    2011-12-01

    The number of microarray and other high-throughput experiments on primary repositories keeps increasing as do the size and complexity of the results in response to biomedical investigations. Initiatives have been started on standardization of content, object model, exchange format and ontology. However, there are backlogs and inability to exchange data between microarray repositories, which indicate that there is a great need for a standard format and data management. We have introduced a metadata framework that includes a metadata card and semantic nets that make experimental results visible, understandable and usable. These are encoded in syntax encoding schemes and represented in RDF (Resource Description Frame-word), can be integrated with other metadata cards and semantic nets, and can be exchanged, shared and queried. We demonstrated the performance and potential benefits through a case study on a selected microarray repository. We concluded that the backlogs can be reduced and that exchange of information and asking of knowledge discovery questions can become possible with the use of this metadata framework.

  6. Damage-plasticity model of the host rock in a nuclear waste repository

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Koudelka, Tomáš; Kruis, Jaroslav, E-mail: kruis@fsv.cvut.cz

    The paper describes damage-plasticity model for the modelling of the host rock environment of a nuclear waste repository. Radioactive Waste Repository Authority in Czech Republic assumes the repository to be in a granite rock mass which exhibit anisotropic behaviour where the strength in tension is lower than in compression. In order to describe this phenomenon, the damage-plasticity model is formulated with the help of the Drucker-Prager yield criterion which can be set to capture the compression behaviour while the tensile stress states is described with the help of scalar isotropic damage model. The concept of damage-plasticity model was implemented inmore » the SIFEL finite element code and consequently, the code was used for the simulation of the Äspö Pillar Stability Experiment (APSE) which was performed in order to determine yielding strength under various conditions in similar granite rocks as in Czech Republic. The results from the performed analysis are presented and discussed in the paper.« less

  7. Testimony of Dr. Raul A. Deju, Basalt Waste Isolation Project, before the Subcommittee on Energy Research and Production, Committee on Sceince and Technology, United States House of Representatives, March 2, 1983

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1983-01-01

    Status of the Basalt Waste Isolation Project is given. Three key concerns have been identified that need to be resolved to either confirm or eliminate the basalts as a potential nuclear waste repository host medium. They are: A thorough understanding of the groundwater hydrology beneath the Hanford Site is needed to assure that a repository in basalt will not contribute unacceptable amounts of contaminants to the accessible environment. Our ability to construct a repository shaft and a network of underground tunnels needs to be fully demonstrated through an exploratory shaft program. Our ability to ultimately seal a repository, such thatmore » its integrity and the isolation of the waste are guaranteed, needs to be demonstrated.« less

  8. IEDA Integrated Services: Improving the User Experience for Interdisciplinary Earth Science Research

    NASA Astrophysics Data System (ADS)

    Carter-Orlando, M.; Ferrini, V. L.; Lehnert, K.; Carbotte, S. M.; Richard, S. M.; Morton, J. J.; Shane, N.; Ash, J.; Song, L.

    2017-12-01

    The Interdisciplinary Earth Data Alliance (IEDA) is an NSF-funded data facility that provides data tools and services to support the Ocean, Earth, and Polar Sciences. IEDA systems, developed and maintained primarily by the IEDA partners EarthChem and the Marine Geoscience Data System (MGDS), serve as primary community data collections for global geochemistry and marine geoscience research and support the preservation, discovery, retrieval, and analysis of a wide range of observational field and analytical data types. Individual IEDA systems originated independently and differ from one another in purpose and scope. Some IEDA systems are data repositories (EarthChem Library, Marine Geo-Digital Library), while others are actively maintained data syntheses (GMRT, PetDB, EarthChem Portal, Geochron). Still others are data visualization and analysis tools (GeoMapApp). Although the diversity of IEDA's data types, tools, and services is a major strength and of high value to investigators, it can be a source of confusion. And while much of the data managed in IEDA systems is appropriate for interdisciplinary research, investigators may be unfamiliar with the user interfaces and services of each system, especially if it is not in their primary discipline. This presentation will highlight new ways in which IEDA helps researchers to more efficiently navigate data submission and data access. It will also discuss how IEDA promotes discovery and access within and across its systems, to serve interdisciplinary science while also remaining aware of and responsive to the more specific needs of its disciplinary user communities. The IEDA Data Submission Hub (DaSH), which is currently under development, aspires to streamline the submission process for both the science data contributor and for the repository data curator. Instead of users deciding a priori, which system they should contribute their data to, the DaSH helps route them to the appropriate repository based primarily on data type, and to efficiently gather the necessary documentation for data accession. Similarly, for those looking for data, the IEDA Data Browser provides cross-system browse and discovery of data in a map interface presented in both Mercator and South Polar projections.

  9. The U.S. Army Person-Event Data Environment: A Military-Civilian Big Data Enterprise.

    PubMed

    Vie, Loryana L; Scheier, Lawrence M; Lester, Paul B; Ho, Tiffany E; Labarthe, Darwin R; Seligman, Martin E P

    2015-06-01

    This report describes a groundbreaking military-civilian collaboration that benefits from an Army and Department of Defense (DoD) big data business intelligence platform called the Person-Event Data Environment (PDE). The PDE is a consolidated data repository that contains unclassified but sensitive manpower, training, financial, health, and medical records covering U.S. Army personnel (Active Duty, Reserve, and National Guard), civilian contractors, and military dependents. These unique data assets provide a veridical timeline capturing each soldier's military experience from entry to separation from the armed forces. The PDE was designed to afford unprecedented cost-efficiencies by bringing researchers and military scientists to a single computerized repository rather than porting vast data resources to individual laboratories. With funding from the Robert Wood Johnson Foundation, researchers from the University of Pennsylvania Positive Psychology Center joined forces with the U.S. Army Research Facilitation Laboratory, forming the scientific backbone of the military-civilian collaboration. This unparalleled opportunity was necessitated by a growing need to learn more about relations between psychological and health assets and health outcomes, including healthcare utilization and costs-issues of major importance for both military and civilian population health. The PDE represents more than 100 times the population size and many times the number of linked variables covered by the nation's leading sources of population health data (e.g., the National Health and Nutrition Examination Survey). Following extensive Army vetting procedures, civilian researchers can mine the PDE's trove of information using a suite of statistical packages made available in a Citrix Virtual Desktop. A SharePoint collaboration and governance management environment ensures user compliance with federal and DoD regulations concerning human subjects' protections and also provides a secure portal for multisite collaborations. Taking similarities and differences between military and civilian populations into account, PDE studies can provide much more detailed insight into health-related questions of broad societal concern. Finding ways to make the rich repository of digitized information in the PDE available through military-civilian collaboration can help solve critical medical and behavioral issues affecting the health and well-being of our nations' military and civilian populations.

  10. Enriching text with images and colored light

    NASA Astrophysics Data System (ADS)

    Sekulovski, Dragan; Geleijnse, Gijs; Kater, Bram; Korst, Jan; Pauws, Steffen; Clout, Ramon

    2008-01-01

    We present an unsupervised method to enrich textual applications with relevant images and colors. The images are collected by querying large image repositories and subsequently the colors are computed using image processing. A prototype system based on this method is presented where the method is applied to song lyrics. In combination with a lyrics synchronization algorithm the system produces a rich multimedia experience. In order to identify terms within the text that may be associated with images and colors, we select noun phrases using a part of speech tagger. Large image repositories are queried with these terms. Per term representative colors are extracted using the collected images. Hereto, we either use a histogram-based or a mean shift-based algorithm. The representative color extraction uses the non-uniform distribution of the colors found in the large repositories. The images that are ranked best by the search engine are displayed on a screen, while the extracted representative colors are rendered on controllable lighting devices in the living room. We evaluate our method by comparing the computed colors to standard color representations of a set of English color terms. A second evaluation focuses on the distance in color between a queried term in English and its translation in a foreign language. Based on results from three sets of terms, a measure of suitability of a term for color extraction based on KL Divergence is proposed. Finally, we compare the performance of the algorithm using either the automatically indexed repository of Google Images and the manually annotated Flickr.com. Based on the results of these experiments, we conclude that using the presented method we can compute the relevant color for a term using a large image repository and image processing.

  11. The NSF Arctic Data Center: Leveraging the DataONE Federation to Build a Sustainable Archive for the NSF Arctic Research Community

    NASA Astrophysics Data System (ADS)

    Budden, A. E.; Arzayus, K. M.; Baker-Yeboah, S.; Casey, K. S.; Dozier, J.; Jones, C. S.; Jones, M. B.; Schildhauer, M.; Walker, L.

    2016-12-01

    The newly established NSF Arctic Data Center plays a critical support role in archiving and curating the data and software generated by Arctic researchers from diverse disciplines. The Arctic community, comprising Earth science, archaeology, geography, anthropology, and other social science researchers, are supported through data curation services and domain agnostic tools and infrastructure, ensuring data are accessible in the most transparent and usable way possible. This interoperability across diverse disciplines within the Arctic community facilitates collaborative research and is mirrored by interoperability between the Arctic Data Center infrastructure and other large scale cyberinfrastructure initiatives. The Arctic Data Center leverages the DataONE federation to standardize access to and replication of data and metadata to other repositories, specifically the NOAA's National Centers for Environmental Information (NCEI). This approach promotes long-term preservation of the data and metadata, as well as opening the door for other data repositories to leverage this replication infrastructure with NCEI and other DataONE member repositories. The Arctic Data Center uses rich, detailed metadata following widely recognized standards. Particularly, measurement-level and provenance metadata provide scientists the details necessary to integrate datasets across studies and across repositories while enabling a full understanding of the provenance of data used in the system. The Arctic Data Center gains this deep metadata and provenance support by simply adopting DataONE services, which results in significant efficiency gains by eliminating the need to develop systems de novo. Similarly, the advanced search tool developed by the Knowledge Network for Biocomplexity and extended for data submission by the Arctic Data Center, can be used by other DataONE-compliant repositories without further development. By standardizing interfaces and leveraging the DataONE federation, the Arctic Data Center has advanced rapidly and can itself contribute to raising the capabilities of all members of the federation.

  12. Collaborative Data Publication Utilizing the Open Data Repository's Data Publisher

    NASA Technical Reports Server (NTRS)

    Stone, N.; Lafuente, B.; Bristow, T.; Keller, R. M.; Downs, R. T.; Blake, D.; Fonda, M.; Dateo, C.; Pires, A.

    2017-01-01

    For small communities in multidisciplinary fields such as astrobiology, publishing and sharing data can be challenging. While large, homogenous fields often have repositories and existing data standards, small groups of independent researchers have few options for publishing data that can be utilized within their community. In conjunction with teams at NASA Ames and the University of Arizona, a number of pilot studies are being conducted to assess the needs of these research groups and to guide the software development so that it allows them to publish and share their data collaboratively.

  13. Status of Progress Made Toward Safety Analysis and Technical Site Evaluations for DOE Managed HLW and SNF.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sevougian, S. David; Stein, Emily; Gross, Michael B

    The Spent Fuel and Waste Science and Technology (SFWST) Campaign of the U.S. Department of Energy (DOE) Office of Nuclear Energy (NE) is conducting research and development (R&D) on generic deep geologic disposal systems (i.e., repositories). This report describes specific activities in FY 2016 associated with the development of a Defense Waste Repository (DWR)a for the permanent disposal of a portion of the HLW and SNF derived from national defense and research and development (R&D) activities of the DOE.

  14. Design and Development of an Institutional Repository at the Indian Institute of Technology Kharagpur

    ERIC Educational Resources Information Center

    Sutradhar, B.

    2006-01-01

    Purpose: To describe how an institutional repository (IR) was set up, using open source software, at the Indian Institute of Technology (IIT) in Kharagpur. Members of the IIT can publish their research documents in the IR for online access as well as digital preservation. Material in this IR includes instructional materials, records, data sets,…

  15. The VLAB OER Experience: Modeling Potential-Adopter Student Acceptance

    ERIC Educational Resources Information Center

    Raman, Raghu; Achuthan, Krishnashree; Nedungadi, Prema; Diwakar, Shyam; Bose, Ranjan

    2014-01-01

    Virtual Labs (VLAB) is a multi-institutional Open Educational Resources (OER) initiative, exclusively focused on lab experiments for engineering education. This project envisages building a large OER repository, containing over 1650 virtual experiments mapped to the engineering curriculum. The introduction of VLAB is a paradigm shift in an…

  16. A web-based repository of surgical simulator projects.

    PubMed

    Leskovský, Peter; Harders, Matthias; Székely, Gábor

    2006-01-01

    The use of computer-based surgical simulators for training of prospective surgeons has been a topic of research for more than a decade. As a result, a large number of academic projects have been carried out, and a growing number of commercial products are available on the market. Keeping track of all these endeavors for established groups as well as for newly started projects can be quite arduous. Gathering information on existing methods, already traveled research paths, and problems encountered is a time consuming task. To alleviate this situation, we have established a modifiable online repository of existing projects. It contains detailed information about a large number of simulator projects gathered from web pages, papers and personal communication. The database is modifiable (with password protected sections) and also allows for a simple statistical analysis of the collected data. For further information, the surgical repository web page can be found at www.virtualsurgery.vision.ee.ethz.ch.

  17. Colloid-Facilitated Radionuclide Transport: Current State of Knowledge from a Nuclear Waste Repository Risk Assessment Perspective

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Reimus, Paul William; Zavarin, Mavrik; Wang, Yifeng

    2017-01-25

    This report provides an overview of the current state of knowledge of colloid-facilitated radionuclide transport from a nuclear waste repository risk assessment perspective. It draws on work that has been conducted over the past 3 decades, although there is considerable emphasis given to work that has been performed over the past 3-5 years as part of the DOE Used Fuel Disposition Campaign. The timing of this report coincides with the completion of a 3-year DOE membership in the Colloids Formation and Migration (CFM) partnership, an international collaboration of scientists studying colloid-facilitated transport of radionuclides at both the laboratory and field-scalesmore » in a fractured crystalline granodiorite at the Grimsel Test Site in Switzerland. This Underground Research Laboratory has hosted the most extensive and carefully-controlled set of colloid-facilitated solute transport experiments that have ever been conducted in an in-situ setting, and a summary of the results to date from these efforts, as they relate to transport over long time and distance scales, is provided in Chapter 3 of this report.« less

  18. Uranium (VI) solubility in carbonate-free ERDA-6 brine

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lucchini, Jean-francois; Khaing, Hnin; Reed, Donald T

    2010-01-01

    When present, uranium is usually an element of importance in a nuclear waste repository. In the Waste Isolation Pilot Plant (WIPP), uranium is the most prevalent actinide component by mass, with about 647 metric tons to be placed in the repository. Therefore, the chemistry of uranium, and especially its solubility in the WIPP conditions, needs to be well determined. Long-term experiments were performed to measure the solubility of uranium (VI) in carbonate-free ERDA-6 brine, a simulated WIPP brine, at pC{sub H+} values between 8 and 12.5. These data, obtained from the over-saturation approach, were the first repository-relevant data for themore » VI actinide oxidation state. The solubility trends observed pointed towards low uranium solubility in WIPP brines and a lack of amphotericity. At the expected pC{sub H+} in the WIPP ({approx} 9.5), measured uranium solubility approached 10{sup -7} M. The objective of these experiments was to establish a baseline solubility to further investigate the effects of carbonate complexation on uranium solubility in WIPP brines.« less

  19. Rolling Deck to Repository (R2R): Standards and Semantics for Open Access to Research Data

    NASA Astrophysics Data System (ADS)

    Arko, Robert; Carbotte, Suzanne; Chandler, Cynthia; Smith, Shawn; Stocks, Karen

    2015-04-01

    In recent years, a growing number of funding agencies and professional societies have issued policies calling for open access to research data. The Rolling Deck to Repository (R2R) program is working to ensure open access to the environmental sensor data routinely acquired by the U.S. academic research fleet. Currently 25 vessels deliver 7 terabytes of data to R2R each year, acquired from a suite of geophysical, oceanographic, meteorological, and navigational sensors on over 400 cruises worldwide. R2R is working to ensure these data are preserved in trusted repositories, discoverable via standard protocols, and adequately documented for reuse. R2R maintains a master catalog of cruises for the U.S. academic research fleet, currently holding essential documentation for over 3,800 expeditions including vessel and cruise identifiers, start/end dates and ports, project titles and funding awards, science parties, dataset inventories with instrument types and file formats, data quality assessments, and links to related content at other repositories. A Digital Object Identifier (DOI) is published for 1) each cruise, 2) each original field sensor dataset, 3) each post-field data product such as quality-controlled shiptrack navigation produced by the R2R program, and 4) each document such as a cruise report submitted by the science party. Scientists are linked to personal identifiers, such as the Open Researcher and Contributor ID (ORCID), where known. Using standard global identifiers such as DOIs and ORCIDs facilitates linking with journal publications and generation of citation metrics. Since its inception, the R2R program has worked in close collaboration with other data repositories in the development of shared semantics for oceanographic research. The R2R cruise catalog uses community-standard terms and definitions hosted by the NERC Vocabulary Server, and publishes ISO metadata records for each cruise that use community-standard profiles developed with the NOAA Data Centers and the EU SeaDataNet project. R2R is a partner in the Ocean Data Interoperability Platform (ODIP), working to strengthen links among regional and national data systems, as well as a lead partner in the EarthCube "GeoLink" project, developing a standard set of ontology design patterns for publishing research data using Semantic Web protocols.

  20. [Radiobiological Human Tissue repository: progress and perspectives for solving the problems of radiation safety and health protection of personnel and population].

    PubMed

    Kirillova, E N; Romanov, S A; Loffredo, C A; Zakharova, M L; Revina, V S; Sokolova, S N; Goerlitz, D S; Zubkova, O V; Lukianova, T V; Uriadnitzkaia, T I; Pavlova, O S; Slukinova, U V; Kolosova, A V; Muksinova, K N

    2014-01-01

    Radiobiological Human Tissue repository was established in order to obtain and store biological material from Mayak PA workers occupationally exposed to ionizing (α- and/or γ-) radiation in a wide dose range, from the residents exposed to long term radiation due to radiation accidents and transfer of the samples to scientists for the purpose of studying the effects of radiation for people and their offspring. The accumulated biomaterial is the informational and research potential that form the basis for the work of the scientists in different spheres of biology and medicine. The repository comprises 5 sections: tumor and non-tumor tissues obtained in the course of autopsies, biopsies, surgeries, samples of blood and its components, of DNA, induced sputum, saliva, and other from people exposed or unexposed (control) to radiation. The biomaterial is stored in formalin, in paraffin blocks, slides, as well as in the freezers under low temperatures. All the information on the samples and the registrants (medical, dosimetry, demographic, and occupational data) was obtained and entered into the electronic database. A constantly updated website of the repository was developed in order to provide a possibility to get acquainted with the material and proceed with application for biosamples for scientists from Russia and abroad. Some data obtained in the course of scientific research works on the basis of the biomaterial from the Repository are briefly introduced in the review.

  1. A Research Graph dataset for connecting research data repositories using RD-Switchboard.

    PubMed

    Aryani, Amir; Poblet, Marta; Unsworth, Kathryn; Wang, Jingbo; Evans, Ben; Devaraju, Anusuriya; Hausstein, Brigitte; Klas, Claus-Peter; Zapilko, Benjamin; Kaplun, Samuele

    2018-05-29

    This paper describes the open access graph dataset that shows the connections between Dryad, CERN, ANDS and other international data repositories to publications and grants across multiple research data infrastructures. The graph dataset was created using the Research Graph data model and the Research Data Switchboard (RD-Switchboard), a collaborative project by the Research Data Alliance DDRI Working Group (DDRI WG) with the aim to discover and connect the related research datasets based on publication co-authorship or jointly funded grants. The graph dataset allows researchers to trace and follow the paths to understanding a body of work. By mapping the links between research datasets and related resources, the graph dataset improves both their discovery and visibility, while avoiding duplicate efforts in data creation. Ultimately, the linked datasets may spur novel ideas, facilitate reproducibility and re-use in new applications, stimulate combinatorial creativity, and foster collaborations across institutions.

  2. Childhood Vesicoureteral Reflux Studies: Registries and Repositories Sources and Nosology

    PubMed Central

    Chesney, Russell W.; Patters, Andrea B.

    2012-01-01

    Despite several recent studies, the advisability of antimicrobial prophylaxis and certain imaging studies for urinary tract infections (UTIs) remains controversial. The role of vesicoureteral reflux (VUR) on the severity and re-infection rates for UTIs is also difficult to assess. Registries and repositories of data and biomaterials from clinical studies in children with VUR are valuable. Disease registries are collections of secondary data related to patients with a specific diagnosis, condition or procedure. Registries differ from indices in that they contain more extensive data. A research repository is an entity that receives, stores, processes and/or disseminates specimens (or other materials) as needed. It encompasses the physical location as well as the full range of activities associated with its operation. It may also be referred to as a biorepository. This report provides information about some current registries and repositories that include data and samples from children with VUR. It also describes the heterogeneous nature of the subjects, as some registries and repositories include only data or samples from patients with primary reflux while others also include those from patients with syndromic or secondary reflux. PMID:23044377

  3. Developing criteria to establish Trusted Digital Repositories

    USGS Publications Warehouse

    Faundeen, John L.

    2017-01-01

    This paper details the drivers, methods, and outcomes of the U.S. Geological Survey’s quest to establish criteria by which to judge its own digital preservation resources as Trusted Digital Repositories. Drivers included recent U.S. legislation focused on data and asset management conducted by federal agencies spending $100M USD or more annually on research activities. The methods entailed seeking existing evaluation criteria from national and international organizations such as International Standards Organization (ISO), U.S. Library of Congress, and Data Seal of Approval upon which to model USGS repository evaluations. Certification, complexity, cost, and usability of existing evaluation models were key considerations. The selected evaluation method was derived to allow the repository evaluation process to be transparent, understandable, and defensible; factors that are critical for judging competing, internal units. Implementing the chosen evaluation criteria involved establishing a cross-agency, multi-disciplinary team that interfaced across the organization. 

  4. Information Warehouse – A Comprehensive Informatics Platform for Business, Clinical, and Research Applications

    PubMed Central

    Kamal, Jyoti; Liu, Jianhua; Ostrander, Michael; Santangelo, Jennifer; Dyta, Ravi; Rogers, Patrick; Mekhjian, Hagop S.

    2010-01-01

    Since its inception in 1997, the IW (Information Warehouse) at the Ohio State University Medical Center (OSUMC) has gradually transformed itself from a single purpose business decision support system to a comprehensive informatics platform supporting basic, clinical, and translational research. The IW today is the combination of four integrated components: a clinical data repository containing over a million patients; a research data repository housing various research specific data; an application development platform for building business and research enabling applications; a business intelligence environment assisting in reporting in all function areas. The IW is structured and encoded using standard terminologies such as SNOMED-CT, ICD, and CPT. The IW is an important component of OSUMC’s Clinical and Translational Science Award (CTSA) informatics program. PMID:21347019

  5. LungMAP: The Molecular Atlas of Lung Development Program

    PubMed Central

    Ardini-Poleske, Maryanne E.; Ansong, Charles; Carson, James P.; Corley, Richard A.; Deutsch, Gail H.; Hagood, James S.; Kaminski, Naftali; Mariani, Thomas J.; Potter, Steven S.; Pryhuber, Gloria S.; Warburton, David; Whitsett, Jeffrey A.; Palmer, Scott M.; Ambalavanan, Namasivayam

    2017-01-01

    The National Heart, Lung, and Blood Institute is funding an effort to create a molecular atlas of the developing lung (LungMAP) to serve as a research resource and public education tool. The lung is a complex organ with lengthy development time driven by interactive gene networks and dynamic cross talk among multiple cell types to control and coordinate lineage specification, cell proliferation, differentiation, migration, morphogenesis, and injury repair. A better understanding of the processes that regulate lung development, particularly alveologenesis, will have a significant impact on survival rates for premature infants born with incomplete lung development and will facilitate lung injury repair and regeneration in adults. A consortium of four research centers, a data coordinating center, and a human tissue repository provides high-quality molecular data of developing human and mouse lungs. LungMAP includes mouse and human data for cross correlation of developmental processes across species. LungMAP is generating foundational data and analysis, creating a web portal for presentation of results and public sharing of data sets, establishing a repository of young human lung tissues obtained through organ donor organizations, and developing a comprehensive lung ontology that incorporates the latest findings of the consortium. The LungMAP website (www.lungmap.net) currently contains more than 6,000 high-resolution lung images and transcriptomic, proteomic, and lipidomic human and mouse data and provides scientific information to stimulate interest in research careers for young audiences. This paper presents a brief description of research conducted by the consortium, database, and portal development and upcoming features that will enhance the LungMAP experience for a community of users. PMID:28798251

  6. MGDB: crossing the marker genes of a user microarray with a database of public-microarrays marker genes.

    PubMed

    Huerta, Mario; Munyi, Marc; Expósito, David; Querol, Enric; Cedano, Juan

    2014-06-15

    The microarrays performed by scientific teams grow exponentially. These microarray data could be useful for researchers around the world, but unfortunately they are underused. To fully exploit these data, it is necessary (i) to extract these data from a repository of the high-throughput gene expression data like Gene Expression Omnibus (GEO) and (ii) to make the data from different microarrays comparable with tools easy to use for scientists. We have developed these two solutions in our server, implementing a database of microarray marker genes (Marker Genes Data Base). This database contains the marker genes of all GEO microarray datasets and it is updated monthly with the new microarrays from GEO. Thus, researchers can see whether the marker genes of their microarray are marker genes in other microarrays in the database, expanding the analysis of their microarray to the rest of the public microarrays. This solution helps not only to corroborate the conclusions regarding a researcher's microarray but also to identify the phenotype of different subsets of individuals under investigation, to frame the results with microarray experiments from other species, pathologies or tissues, to search for drugs that promote the transition between the studied phenotypes, to detect undesirable side effects of the treatment applied, etc. Thus, the researcher can quickly add relevant information to his/her studies from all of the previous analyses performed in other studies as long as they have been deposited in public repositories. Marker-gene database tool: http://ibb.uab.es/mgdb © The Author 2014. Published by Oxford University Press.

  7. Modelling the Mont Terri HE-D experiment for the Thermal–Hydraulic–Mechanical response of a bedded argillaceous formation to heating

    DOE PAGES

    Garitte, B.; Nguyen, T. S.; Barnichon, J. D.; ...

    2017-05-09

    Coupled thermal–hydrological–mechanical (THM) processes in the near field of deep geological repositories can influence several safety features of the engineered and geological barriers. Among those features are: the possibility of damage in the host rock, the time for re-saturation of the bentonite, and the perturbations in the hydraulic regime in both the rock and engineered seals. Within the international cooperative code-validation project DECOVALEX-2015, eight research teams developed models to simulate an in situ heater experiment, called HE-D, in Opalinus Clay at the Mont Terri Underground Research Laboratory in Switzerland. The models were developed from the theory of poroelasticity in ordermore » to simulate the coupled THM processes that prevailed during the experiment and thereby to characterize the in situ THM properties of Opalinus Clay. The modelling results for the evolution of temperature, pore water pressure, and deformation at different points are consistent among the research teams and compare favourably with the experimental data in terms of trends and absolute values. The models were able to reproduce the main physical processes of the experiment. In particular, most teams simulated temperature and thermally induced pore water pressure well, including spatial variations caused by inherent anisotropy due to bedding.« less

  8. Modelling the Mont Terri HE-D experiment for the Thermal–Hydraulic–Mechanical response of a bedded argillaceous formation to heating

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Garitte, B.; Nguyen, T. S.; Barnichon, J. D.

    Coupled thermal–hydrological–mechanical (THM) processes in the near field of deep geological repositories can influence several safety features of the engineered and geological barriers. Among those features are: the possibility of damage in the host rock, the time for re-saturation of the bentonite, and the perturbations in the hydraulic regime in both the rock and engineered seals. Within the international cooperative code-validation project DECOVALEX-2015, eight research teams developed models to simulate an in situ heater experiment, called HE-D, in Opalinus Clay at the Mont Terri Underground Research Laboratory in Switzerland. The models were developed from the theory of poroelasticity in ordermore » to simulate the coupled THM processes that prevailed during the experiment and thereby to characterize the in situ THM properties of Opalinus Clay. The modelling results for the evolution of temperature, pore water pressure, and deformation at different points are consistent among the research teams and compare favourably with the experimental data in terms of trends and absolute values. The models were able to reproduce the main physical processes of the experiment. In particular, most teams simulated temperature and thermally induced pore water pressure well, including spatial variations caused by inherent anisotropy due to bedding.« less

  9. Arpeggio: harmonic compression of ChIP-seq data reveals protein-chromatin interaction signatures

    PubMed Central

    Stanton, Kelly Patrick; Parisi, Fabio; Strino, Francesco; Rabin, Neta; Asp, Patrik; Kluger, Yuval

    2013-01-01

    Researchers generating new genome-wide data in an exploratory sequencing study can gain biological insights by comparing their data with well-annotated data sets possessing similar genomic patterns. Data compression techniques are needed for efficient comparisons of a new genomic experiment with large repositories of publicly available profiles. Furthermore, data representations that allow comparisons of genomic signals from different platforms and across species enhance our ability to leverage these large repositories. Here, we present a signal processing approach that characterizes protein–chromatin interaction patterns at length scales of several kilobases. This allows us to efficiently compare numerous chromatin-immunoprecipitation sequencing (ChIP-seq) data sets consisting of many types of DNA-binding proteins collected from a variety of cells, conditions and organisms. Importantly, these interaction patterns broadly reflect the biological properties of the binding events. To generate these profiles, termed Arpeggio profiles, we applied harmonic deconvolution techniques to the autocorrelation profiles of the ChIP-seq signals. We used 806 publicly available ChIP-seq experiments and showed that Arpeggio profiles with similar spectral densities shared biological properties. Arpeggio profiles of ChIP-seq data sets revealed characteristics that are not easily detected by standard peak finders. They also allowed us to relate sequencing data sets from different genomes, experimental platforms and protocols. Arpeggio is freely available at http://sourceforge.net/p/arpeggio/wiki/Home/. PMID:23873955

  10. Arpeggio: harmonic compression of ChIP-seq data reveals protein-chromatin interaction signatures.

    PubMed

    Stanton, Kelly Patrick; Parisi, Fabio; Strino, Francesco; Rabin, Neta; Asp, Patrik; Kluger, Yuval

    2013-09-01

    Researchers generating new genome-wide data in an exploratory sequencing study can gain biological insights by comparing their data with well-annotated data sets possessing similar genomic patterns. Data compression techniques are needed for efficient comparisons of a new genomic experiment with large repositories of publicly available profiles. Furthermore, data representations that allow comparisons of genomic signals from different platforms and across species enhance our ability to leverage these large repositories. Here, we present a signal processing approach that characterizes protein-chromatin interaction patterns at length scales of several kilobases. This allows us to efficiently compare numerous chromatin-immunoprecipitation sequencing (ChIP-seq) data sets consisting of many types of DNA-binding proteins collected from a variety of cells, conditions and organisms. Importantly, these interaction patterns broadly reflect the biological properties of the binding events. To generate these profiles, termed Arpeggio profiles, we applied harmonic deconvolution techniques to the autocorrelation profiles of the ChIP-seq signals. We used 806 publicly available ChIP-seq experiments and showed that Arpeggio profiles with similar spectral densities shared biological properties. Arpeggio profiles of ChIP-seq data sets revealed characteristics that are not easily detected by standard peak finders. They also allowed us to relate sequencing data sets from different genomes, experimental platforms and protocols. Arpeggio is freely available at http://sourceforge.net/p/arpeggio/wiki/Home/.

  11. Multi-institutional tumor banking: lessons learned from a pancreatic cancer biospecimen repository.

    PubMed

    Demeure, Michael J; Sielaff, Timothy; Koep, Larry; Prinz, Richard; Moser, A James; Zeh, Herb; Hostetter, Galen; Black, Jodi; Decker, Ardis; Rosewell, Sandra; Bussey, Kimberly J; Von Hoff, Daniel

    2010-10-01

    Clinically annotated pancreatic cancer samples are needed for progress to be made toward developing more effective treatments for this deadly cancer. As part of a National Cancer Institute-funded program project, we established a biospecimen core to support the research efforts. This article summarizes the key hurdles encountered and solutions we found in the process of developing a successful multi-institution biospecimen repository.

  12. Linking User Identities Across the DataONE Federation of Data Repositories

    NASA Astrophysics Data System (ADS)

    Jones, M. B.; Mecum, B.; Leinfelder, B.; Jones, C. S.; Walker, L.

    2016-12-01

    DataONE provides services for identifying, authenticating, and authorizing researchers to access and contribute data to repositories within the DataONE federation. In the earth sciences, thousands of institutional and disciplinary repositories have created their own user identity and authentication systems with their own user directory based on a database or web content management systems. Thus, researchers have many identities that are neither linked nor interoperable, making it difficult to reference the identity of these users across systems. Key user information is hidden, and only a non-disambiguated name is often available. From a sample of 160,000 data sets within DataONE, a super-majority of references to the data creators lack even an email address. In an attempt to disambiguate these people via the GeoLink project, we conservatively estimate they represent at least 57,000 unique identities, but without a clear user identifier, there could be as many as 223,000. Interoperability among repositories is critical to improving the scope of scientific synthesis and capabilities for research collaboration. While many have focused on the convenience of Single Sign-On (SSO), we have found that sharing user identifiers is far more useful for interoperability. With an unambiguous user identity in incoming metadata, DataONE has built user-profiles that present that user's data across repositories, that link users and their organizational affiliations, and that allow users to work collaboratively in private groups that span repository systems. DataONE's user identity solution leverages existing systems such as InCommon, CILogon, Google, and ORCID to not further proliferate user identities. DataONE provides a core service allowing users to link their multiple identities so that authenticating with one identity (e.g., ORCID) can authorize access to data protected via another identity (e.g., InCommon). Currently, DataONE is using ORCID identities to link and identify users, but challenges must still be overcome to support historical records for which ORCIDs can not be used because the associated people are unavailable to confirm their identity. DataONE's identity systems facilitate crosslinking between user identities and scientific metadata to accelerate collaboration and synthesis.

  13. Smoothing Data Friction through building Service Oriented Data Platforms

    NASA Astrophysics Data System (ADS)

    Wyborn, L. A.; Richards, C. J.; Evans, B. J. K.; Wang, J.; Druken, K. A.

    2017-12-01

    Data Friction has been commonly defined as the costs in time, energy and attention required to simply collect, check, store, move, receive, and access data. On average, researchers spend a significant fraction of their time finding the data for their research project and then reformatting it so that it can be used by the software application of their choice. There is an increasing role for both data repositories and software to be modernised to help reduce data friction in ways that support the better use of the data. Many generic data repositories simply accept data in the format as supplied: the key check is that the data have sufficient metadata to enable discovery and download. Few generic repositories have both the expertise and infrastructure to support the multiple domain specific requirements that facilitate the increasing need for integration and reusability. In contrast, major science domain-focused repositories are increasingly able to implement and enforce community endorsed best practices and guidelines that ensure reusability and harmonization of data for use within the community by offering semi-automated QC workflows to improve quality of submitted data. The most advanced of these science repositories now operate as service-oriented data platforms that extend the use of data across domain silos and increasingly provide server-side programmatically-enabled access to data via network protocols and community standard APIs. To provide this, more rigorous QA/QC procedures are needed to validate data against standards and community software and tools. This ensures that the data can be accessed in expected ways and also demonstrates that the data works across different (non-domain specific) packages, tools and programming languages deployed by the various user communities. In Australia, the National Computational Infrastructure (NCI) has created such a service-oriented data platform which is demonstrating how this approach can reduce data friction, servicing both individual domains as well as facilitating cross-domain collaboration. The approach has required an increase in effort for the repository to provide the additional expertise, so as to enable a better capability and efficient system which ultimately saves time by the individual researcher.

  14. Quantification of Cation Sorption to Engineered Barrier Materials Under Extreme Conditions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Powell, Brian; Schlautman, Mark; Rao, Linfeng

    The objective of this research is to examine mechanisms and thermodynamics of actinide sorption to engineered barrier materials (iron (oxyhydr)oxides and bentonite clay) for nuclear waste repositories under high temperature and high ionic strength conditions using a suite of macroscopic and microscopic techniques which will be coupled with interfacial reaction models. Gaining a mechanistic understanding of interfacial processes governing the sorption/sequestration of actinides at mineral-water interfaces is fundamental for the accurate prediction of actinide behavior in waste repositories. Although macroscale sorption data and various spectroscopic techniques have provided valuable information regarding speciation of actinides at solid-water interfaces, significant knowledge gapsmore » still exist with respect to sorption mechanisms and the ability to quantify sorption, particularly at high temperatures and ionic strengths. This objective is addressed through three major tasks: (1) influence of oxidation state on actinide sorption to iron oxides and clay minerals at elevated temperatures and ionic strengths; (2) calorimetric titrations of actinide-mineral suspensions; (3) evaluation of bentonite performance under repository conditions. The results of the work will include a qualitative conceptual model and a quantitative thermodynamic speciation model describing actinide partitioning to minerals and sediments, which is based upon a mechanistic understanding of specific sorption processes as determined from both micro-scale and macroscale experimental techniques. The speciation model will be a thermodynamic aqueous and surface complexation model of actinide interactions with mineral surfaces that is self-consistent with macroscopic batch sorption data, calorimetric and potentiometric titrations, X-ray absorption Spectroscopy (XAS, mainly Extended X-ray Absorption Fine Structure (EXAFS)), and electron microscopy analyses. The novelty of the proposed work lies largely in the unique system conditions which will be examined (i.e. elevated temperature and ionic strength) and the manner in which the surface complexation model will be developed in terms of specific surface species identified using XAS. These experiments will thus provide a fundamental understanding of the chemical and physical processes occurring at the solid-solution interface under expected repository conditions. Additionally, the focus on thermodynamic treatment of actinide ion interactions with minerals as proposed will provide information on the driving forces involved and contribute to the overall understanding of the high affinity many actinide ions have for oxide surfaces. The utility of this model will be demonstrated in this work through a series of advective and diffusive flow experiments.« less

  15. The beliefs, motivations, and expectations of parents who have enrolled their children in a genetic biorepository.

    PubMed

    Harris, Erin D; Ziniel, Sonja I; Amatruda, Jonathan G; Clinton, Catherine M; Savage, Sarah K; Taylor, Patrick L; Huntington, Noelle L; Green, Robert C; Holm, Ingrid A

    2012-03-01

    Little is known about parental attitudes toward return of individual research results (IRRs) in pediatric genomic research. The aim of this study was to understand the views of the parents who enrolled their children in a genomic repository in which IRRs will be returned. We conducted focus groups with parents of children with developmental disorders enrolled in the Gene Partnership (GP), a genomic research repository that offers to return IRRs, to learn about their understanding of the GP, motivations for enrolling their children, and expectations regarding the return of IRRs. Parents hoped to receive IRRs that would help them better understand their children's condition(s). They understood that this outcome was unlikely, but hoped that their children's participation in the GP would contribute to scientific knowledge. Most parents wanted to receive all IRRs about their child, even for diseases that were severe and untreatable, citing reasons of personal utility. Parents preferred electronic delivery of the results and wanted to designate their preferences regarding what information they would receive. It is important for researchers to understand participant expectations in enrolling in a research repository that offers to disclose children's IRRs in order to effectively communicate the implications to parents during the consenting process.

  16. The visualization and availability of experimental research data at Elsevier

    NASA Astrophysics Data System (ADS)

    Keall, Bethan

    2014-05-01

    In the digital age, the visualization and availability of experimental research data is an increasingly prominent aspect of the research process and of the scientific output that researchers generate. We expect that the importance of data will continue to grow, driven by technological advancements, requirements from funding bodies to make research data available, and a developing research data infrastructure that is supported by data repositories, science publishers, and other stakeholders. Elsevier is actively contributing to these efforts, for example by setting up bidirectional links between online articles on ScienceDirect and relevant data sets on trusted data repositories. A key aspect of Elsevier's "Article of the Future" program, these links enrich the online article and make it easier for researchers to find relevant data and articles and help place data in the right context for re-use. Recently, we have set up such links with some of the leading data repositories in Earth Sciences, including the British Geological Survey, Integrated Earth Data Applications, the UK Natural Environment Research Council, and the Oak Ridge National Laboratory DAAC. Building on these links, Elsevier has also developed a number of data integration and visualization tools, such as an interactive map viewer that displays the locations of relevant data from PANGAEA next to articles on ScienceDirect. In this presentation we will give an overview of these and other capabilities of the Article of the Future, focusing on how they help advance communication of research in the digital age.

  17. The Full Scale Seal Experiment - A Seal Industrial Prototype for Cigeo - 13106

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lebon, P.; Bosgiraud, J.M.; Foin, R.

    2013-07-01

    The Full Scale Seal (FSS) Experiment is one of various experiments implemented by Andra, within the frame of the Cigeo (the French Deep Geological Repository) Project development, to demonstrate the technical construction feasibility and performance of seals to be constructed, at time of Repository components (shafts, ramps, drifts, disposal vaults) progressive closure. FSS is built inside a drift model fabricated on surface for the purpose. Prior to the scale 1:1 seal construction test, various design tasks are scheduled. They include the engineering work on the drift model to make it fit with the experimental needs, on the various work sequencesmore » anticipated for the swelling clay core emplacement and the concrete containment plugs construction, on the specialized handling tools (and installation equipment) manufactured and delivered for the purpose, and of course on the various swelling clay materials and low pH (below 11) concrete formulations developed for the application. The engineering of the 'seal-as-built' commissioning means (tools and methodology) must also be dealt with. The FSS construction experiment is a technological demonstrator, thus it is not focused on the phenomenological survey (and by consequence, on the performance and behaviour forecast). As such, no hydration (forced or natural) is planned. However, the FSS implementation (in particular via the construction and commissioning activities carried out) is a key milestone in view of comforting phenomenological extrapolation in time and scale. The FSS experiment also allows for qualifying the commissioning methods of a real sealing system in the Repository, as built, at time of industrial operations. (authors)« less

  18. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Choung, Sungwook; Um, Wooyong; Pacific Northwest National Laboratory

    Permanent disposal of low- and intermediate-level radioactive wastes in the subterranean environment has been the preferred method of many countries, including Korea. A safety issue after the closure of a geological repository is that biodegradation of organic materials due to microbial activities generates gases that lead to overpressure of the waste containers in the repository and its disintegration with the release of radionuclides. As part of an ongoing large-scale in situ experiment using organic wastes and groundwater to simulate geological radioactive waste repository conditions, we investigated the geochemical alteration and microbial activities at an early stage (~63 days) intended tomore » be representative of the initial period after repository closure. The increased numbers of both aerobes and facultative anaerobes in waste effluents indicate that oxygen content could be the most significant parameter to control biogeochemical conditions at very early periods of reaction (<35 days). Accordingly, the values of dissolved oxygen and redox potential were decreased. The activation of anaerobes after 35 days was supported by the increased concentration to ~50 mg L-1 of ethanol. These results suggest that the biogeochemical conditions were rapidly altered to more reducing and anaerobic conditions within the initial 2 months after repository closure. Although no gases were detected during the study, activated anaerobic microbes will play more important role in gas generation over the long term.« less

  19. The Legacy Archive for Microwave Background Data Analysis (LAMBDA)

    NASA Astrophysics Data System (ADS)

    Miller, Nathan; LAMBDA

    2018-01-01

    The Legacy Archive for Microwave Background Data Analysis (LAMBDA) provides CMB researchers with archival data for cosmology missions, software tools, and links to other sites of interest. LAMBDA is one-stop shopping for CMB researchers. It hosts data from WMAP along with many suborbital experiments. Over the past year, LAMBDA has acquired new data from SPTpol, SPIDER and ACTPol. In addition to the primary CMB, LAMBDA also provides foreground data.LAMBDA has several ongoing efforts to provide tools for CMB researchers. These tools include a web interface for CAMB and a web interface for a CMB survey footprint database and plotting tool. Additionally, we have recently developed a Docker container with standard CMB analysis tools and demonstrations in the form of Jupyter notebooks. These containers will be publically available through Docker's container repository and the source will be available on github.

  20. eXframe: reusable framework for storage, analysis and visualization of genomics experiments

    PubMed Central

    2011-01-01

    Background Genome-wide experiments are routinely conducted to measure gene expression, DNA-protein interactions and epigenetic status. Structured metadata for these experiments is imperative for a complete understanding of experimental conditions, to enable consistent data processing and to allow retrieval, comparison, and integration of experimental results. Even though several repositories have been developed for genomics data, only a few provide annotation of samples and assays using controlled vocabularies. Moreover, many of them are tailored for a single type of technology or measurement and do not support the integration of multiple data types. Results We have developed eXframe - a reusable web-based framework for genomics experiments that provides 1) the ability to publish structured data compliant with accepted standards 2) support for multiple data types including microarrays and next generation sequencing 3) query, analysis and visualization integration tools (enabled by consistent processing of the raw data and annotation of samples) and is available as open-source software. We present two case studies where this software is currently being used to build repositories of genomics experiments - one contains data from hematopoietic stem cells and another from Parkinson's disease patients. Conclusion The web-based framework eXframe offers structured annotation of experiments as well as uniform processing and storage of molecular data from microarray and next generation sequencing platforms. The framework allows users to query and integrate information across species, technologies, measurement types and experimental conditions. Our framework is reusable and freely modifiable - other groups or institutions can deploy their own custom web-based repositories based on this software. It is interoperable with the most important data formats in this domain. We hope that other groups will not only use eXframe, but also contribute their own useful modifications. PMID:22103807

  1. Tourism impacts of Three Mile Island and other adverse events: Implications for Lincoln County and other rural counties bisected by radioactive wastes intended for Yucca Mountain

    NASA Astrophysics Data System (ADS)

    Himmelberger, Jeffery J.; Baughman, Mike; Ogneva-Himmelberger, Yelena A.

    1995-11-01

    Whether the proposed Yucca Mountain nuclear waste repository system will adversely impact tourism in southern Nevada is an open question of particular importance to visitor-oriented rural counties bisected by planned waste transportatin corridors (highway or rail). As part of one such county's repository impact assessment program, tourism implications of Three Mile Island (TMI) and other major hazard events have beem revisited to inform ongoing county-wide socioeconomic assessments and contingency planning efforts. This paper summarizes key research implications of such research as applied to Lincoln County, Nevada. Implications for other rural counties are discussed in light of the research findings.

  2. Development of DKB ETL module in case of data conversion

    NASA Astrophysics Data System (ADS)

    Kaida, A. Y.; Golosova, M. V.; Grigorieva, M. A.; Gubin, M. Y.

    2018-05-01

    Modern scientific experiments involve the producing of huge volumes of data that requires new approaches in data processing and storage. These data themselves, as well as their processing and storage, are accompanied by a valuable amount of additional information, called metadata, distributed over multiple informational systems and repositories, and having a complicated, heterogeneous structure. Gathering these metadata for experiments in the field of high energy nuclear physics (HENP) is a complex issue, requiring the quest for solutions outside the box. One of the tasks is to integrate metadata from different repositories into some kind of a central storage. During the integration process, metadata taken from original source repositories go through several processing steps: metadata aggregation, transformation according to the current data model and loading it to the general storage in a standardized form. The R&D project of ATLAS experiment on LHC, Data Knowledge Base, is aimed to provide fast and easy access to significant information about LHC experiments for the scientific community. The data integration subsystem, being developed for the DKB project, can be represented as a number of particular pipelines, arranging data flow from data sources to the main DKB storage. The data transformation process, represented by a single pipeline, can be considered as a number of successive data transformation steps, where each step is implemented as an individual program module. This article outlines the specifics of program modules, used in the dataflow, and describes one of the modules developed and integrated into the data integration subsystem of DKB.

  3. Criteria for the evaluation and certification of long-term digital archives in the earth sciences

    NASA Astrophysics Data System (ADS)

    Klump, Jens

    2010-05-01

    Digital information has become an indispensable part of our cultural and scientific heritage. Scientific findings, historical documents and cultural achievements are to a rapidly increasing extent being presented in electronic form - in many cases exclusively so. However, besides the invaluable advantages offered by this form, it also carries a serious disadvantage: users need to invest a great deal of technical effort in accessing the information. Also, the underlying technology is still undergoing further development at an exceptionally fast pace. The rapid obsolescence of the technology required to read the information combined with the frequently imperceptible physical decay of the media themselves represents a serious threat to preservation of the information content. Many data sets in earth science research are from observations that cannot be repeated. This makes these digital assets particularly valuable. Therefore, these data should be kept and made available for re-use long after the end of the project from which they originated. Since research projects only run for a relatively short period of time, it is advisable to shift the burden of responsibility for long-term data curation from the individual researcher to a trusted data repository or archive. But what makes a trusted data repository? Each trusted digital repository has its own targets and specifications. The trustworthiness of digital repositories can be tested and assessed on the basis of a criteria catalogue. This is the main focus of the work of the nestor working group "Trusted repositories - Certification". It identifies criteria which permit the trustworthiness of a digital repository to be evaluated, both at the organisational and technical levels. The criteria are defined in close collaboration with a wide range of different memory organisations, producers of information, experts and other interested parties. This open approach ensures a high degree of universal validity, suitability for daily practical use and also broad-based acceptance of the results. The criteria catalogue is also intended to present the option of documenting trustworthiness by means of certification in a standardised national or international process. The criteria catalogue is based on the Reference Model for an Open Archival Information System (OAIS, ISO 14721:2003) With its broad approach, the nestor criteria catalogue for trusted digital repositories has to remain on a high level of abstraction. For application in the earth sciences the evaluation criteria need to be transferred into the context of earth science data and their designated user community. This presentation offers a brief introduction to the problems surrounding the long-term preservation of digital objects. This introduction is followed by a proposed application of the criteria catalogue for trusted digital repositories to the context of earth science data and their long-term preservation.

  4. CB Database: A change blindness database for objects in natural indoor scenes.

    PubMed

    Sareen, Preeti; Ehinger, Krista A; Wolfe, Jeremy M

    2016-12-01

    Change blindness has been a topic of interest in cognitive sciences for decades. Change detection experiments are frequently used for studying various research topics such as attention and perception. However, creating change detection stimuli is tedious and there is no open repository of such stimuli using natural scenes. We introduce the Change Blindness (CB) Database with object changes in 130 colored images of natural indoor scenes. The size and eccentricity are provided for all the changes as well as reaction time data from a baseline experiment. In addition, we have two specialized satellite databases that are subsets of the 130 images. In one set, changes are seen in rooms or in mirrors in those rooms (Mirror Change Database). In the other, changes occur in a room or out a window (Window Change Database). Both the sets have controlled background, change size, and eccentricity. The CB Database is intended to provide researchers with a stimulus set of natural scenes with defined stimulus parameters that can be used for a wide range of experiments. The CB Database can be found at http://search.bwh.harvard.edu/new/CBDatabase.html .

  5. A Safety Case Approach for Deep Geologic Disposal of DOE HLW and DOE SNF in Bedded Salt - 13350

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sevougian, S. David; MacKinnon, Robert J.; Leigh, Christi D.

    2013-07-01

    The primary objective of this study is to investigate the feasibility and utility of developing a defensible safety case for disposal of United States Department of Energy (U.S. DOE) high-level waste (HLW) and DOE spent nuclear fuel (SNF) in a conceptual deep geologic repository that is assumed to be located in a bedded salt formation of the Delaware Basin [1]. A safety case is a formal compilation of evidence, analyses, and arguments that substantiate and demonstrate the safety of a proposed or conceptual repository. We conclude that a strong initial safety case for potential licensing can be readily compiled bymore » capitalizing on the extensive technical basis that exists from prior work on the Waste Isolation Pilot Plant (WIPP), other U.S. repository development programs, and the work published through international efforts in salt repository programs such as in Germany. The potential benefits of developing a safety case include leveraging previous investments in WIPP to reduce future new repository costs, enhancing the ability to effectively plan for a repository and its licensing, and possibly expediting a schedule for a repository. A safety case will provide the necessary structure for organizing and synthesizing existing salt repository science and identifying any issues and gaps pertaining to safe disposal of DOE HLW and DOE SNF in bedded salt. The safety case synthesis will help DOE to plan its future R and D activities for investigating salt disposal using a risk-informed approach that prioritizes test activities that include laboratory, field, and underground investigations. It should be emphasized that the DOE has not made any decisions regarding the disposition of DOE HLW and DOE SNF. Furthermore, the safety case discussed herein is not intended to either site a repository in the Delaware Basin or preclude siting in other media at other locations. Rather, this study simply presents an approach for accelerated development of a safety case for a potential DOE HLW and DOE SNF repository using the currently available technical basis for bedded salt. This approach includes a summary of the regulatory environment relevant to disposal of DOE HLW and DOE SNF in a deep geologic repository, the key elements of a safety case, the evolution of the safety case through the successive phases of repository development and licensing, and the existing technical basis that could be used to substantiate the safety of a geologic repository if it were to be sited in the Delaware Basin. We also discuss the potential role of an underground research laboratory (URL). (authors)« less

  6. Selecting a Laboratory Information Management System for Biorepositories in Low- and Middle-Income Countries: The H3Africa Experience and Lessons Learned

    PubMed Central

    Musinguzi, Henry; Lwanga, Newton; Kezimbira, Dafala; Kigozi, Edgar; Katabazi, Fred Ashaba; Wayengera, Misaki; Joloba, Moses Lutaakome; Abayomi, Emmanuel Akin; Swanepoel, Carmen; Croxton, Talishiea; Ozumba, Petronilla; Thankgod, Anazodo; van Zyl, Lizelle; Mayne, Elizabeth Sarah; Kader, Mukthar; Swartz, Garth

    2017-01-01

    Biorepositories in Africa need significant infrastructural support to meet International Society for Biological and Environmental Repositories (ISBER) Best Practices to support population-based genomics research. ISBER recommends a biorepository information management system which can manage workflows from biospecimen receipt to distribution. The H3Africa Initiative set out to develop regional African biorepositories where Uganda, Nigeria, and South Africa were successfully awarded grants to develop the state-of-the-art biorepositories. The biorepositories carried out an elaborate process to evaluate and choose a laboratory information management system (LIMS) with the aim of integrating the three geographically distinct sites. In this article, we review the processes, African experience, lessons learned, and make recommendations for choosing a biorepository LIMS in the African context.

  7. The development of digital library system for drug research information.

    PubMed

    Kim, H J; Kim, S R; Yoo, D S; Lee, S H; Suh, O K; Cho, J H; Shin, H T; Yoon, J P

    1998-01-01

    The sophistication of computer technology and information transmission on internet has made various cyber information repository available to information consumers. In the era of information super-highway, the digital library which can be accessed from remote sites at any time is considered the prototype of information repository. Using object-oriented DBMS, the very first model of digital library for pharmaceutical researchers and related professionals in Korea has been developed. The published research papers and researchers' personal information was included in the database. For database with research papers, 13 domestic journals were abstracted and scanned for full-text image files which can be viewed by Internet web browsers. The database with researchers' personal information was also developed and interlinked to the database with research papers. These database will be continuously updated and will be combined with world-wide information as the unique digital library in the field of pharmacy.

  8. Sharing individual participant data from clinical trials: an opinion survey regarding the establishment of a central repository.

    PubMed

    Tudur Smith, Catrin; Dwan, Kerry; Altman, Douglas G; Clarke, Mike; Riley, Richard; Williamson, Paula R

    2014-01-01

    Calls have been made for increased access to individual participant data (IPD) from clinical trials, to ensure that complete evidence is available. However, despite the obvious benefits, progress towards this is frustratingly slow. In the meantime, many systematic reviews have already collected IPD from clinical trials. We propose that a central repository for these IPD should be established to ensure that these datasets are safeguarded and made available for use by others, building on the strengths and advantages of the collaborative groups that have been brought together in developing the datasets. Evaluate the level of support, and identify major issues, for establishing a central repository of IPD. On-line survey with email reminders. 71 reviewers affiliated with the Cochrane Collaboration's IPD Meta-analysis Methods Group were invited to participate. 30 (42%) invitees responded: 28 (93%) had been involved in an IPD review and 24 (80%) had been involved in a randomised trial. 25 (83%) agreed that a central repository was a good idea and 25 (83%) agreed that they would provide their IPD for central storage. Several benefits of a central repository were noted: safeguarding and standardisation of data, increased efficiency of IPD meta-analyses, knowledge advancement, and facilitating future clinical, and methodological research. The main concerns were gaining permission from trial data owners, uncertainty about the purpose of the repository, potential resource implications, and increased workload for IPD reviewers. Restricted access requiring approval, data security, anonymisation of data, and oversight committees were highlighted as issues under governance of the repository. There is support in this community of IPD reviewers, many of whom are also involved in clinical trials, for storing IPD in a central repository. Results from this survey are informing further work on developing a repository of IPD which is currently underway by our group.

  9. Shared Medical Imaging Repositories.

    PubMed

    Lebre, Rui; Bastião, Luís; Costa, Carlos

    2018-01-01

    This article describes the implementation of a solution for the integration of ownership concept and access control over medical imaging resources, making possible the centralization of multiple instances of repositories. The proposed architecture allows the association of permissions to repository resources and delegation of rights to third entities. It includes a programmatic interface for management of proposed services, made available through web services, with the ability to create, read, update and remove all components resulting from the architecture. The resulting work is a role-based access control mechanism that was integrated with Dicoogle Open-Source Project. The solution has several application scenarios like, for instance, collaborative platforms for research and tele-radiology services deployed at Cloud.

  10. Experiments with Analytic Centers: A confluence of data, tools and help in using them.

    NASA Astrophysics Data System (ADS)

    Little, M. M.; Crichton, D. J.; Hines, K.; Cole, M.; Quam, B. M.

    2017-12-01

    Traditional repositories have been primarily focused on data stewardship. Over the past two decades, data scientists have attempted to overlay a superstructure to make these repositories more amenable to analysis tasks, with limited success. This poster will summarize lessons learned and some realizations regarding what it takes to create an analytic center. As the volume of Earth Science data grows and the sophistication of analytic tools improves, a pattern has emerged that indicates different science communities uniquely apply a selection of tools to the data to produce scientific results. Infrequently do the experiences of one group help steer other groups. How can the information technology community seed these domains with tools that conform to the thought processes and experiences of that particular science group? What types of succcessful technology infusions have occured and how does technology get adopted. AIST has been experimenting with the management of this analytic center process; this paper will summarize the results and indicate a direction for future infusion attempts.

  11. An overview of the U.S. Army Research Laboratory's Sensor Information Testbed for Collaborative Research Environment (SITCORE) and Automated Online Data Repository (AODR) capabilities

    NASA Astrophysics Data System (ADS)

    Ward, Dennis W.; Bennett, Kelly W.

    2017-05-01

    The Sensor Information Testbed COllaberative Research Environment (SITCORE) and the Automated Online Data Repository (AODR) are significant enablers of the U.S. Army Research Laboratory (ARL)'s Open Campus Initiative and together create a highly-collaborative research laboratory and testbed environment focused on sensor data and information fusion. SITCORE creates a virtual research development environment allowing collaboration from other locations, including DoD, industry, academia, and collation facilities. SITCORE combined with AODR provides end-toend algorithm development, experimentation, demonstration, and validation. The AODR enterprise allows the U.S. Army Research Laboratory (ARL), as well as other government organizations, industry, and academia to store and disseminate multiple intelligence (Multi-INT) datasets collected at field exercises and demonstrations, and to facilitate research and development (R and D), and advancement of analytical tools and algorithms supporting the Intelligence, Surveillance, and Reconnaissance (ISR) community. The AODR provides a potential central repository for standards compliant datasets to serve as the "go-to" location for lessons-learned and reference products. Many of the AODR datasets have associated ground truth and other metadata which provides a rich and robust data suite for researchers to develop, test, and refine their algorithms. Researchers download the test data to their own environments using a sophisticated web interface. The AODR allows researchers to request copies of stored datasets and for the government to process the requests and approvals in an automated fashion. Access to the AODR requires two-factor authentication in the form of a Common Access Card (CAC) or External Certificate Authority (ECA)

  12. Modelling of processes occurring in deep geological repository - Development of new modules in the GoldSim environment

    NASA Astrophysics Data System (ADS)

    Vopálka, D.; Lukin, D.; Vokál, A.

    2006-01-01

    Three new modules modelling the processes that occur in a deep geological repository have been prepared in the GoldSim computer code environment (using its Transport Module). These modules help to understand the role of selected parameters in the near-field region of the final repository and to prepare an own complex model of the repository behaviour. The source term module includes radioactive decay and ingrowth in the canister, first order degradation of fuel matrix, solubility limitation of the concentration of the studied nuclides, and diffusive migration through the surrounding bentonite layer controlled by the output boundary condition formulated with respect to the rate of water flow in the rock. The corrosion module describes corrosion of canisters made of carbon steel and transport of corrosion products in the near-field region. This module computes balance equations between dissolving species and species transported by diffusion and/or advection from the surface of a solid material. The diffusion module that includes also non-linear form of the interaction isotherm can be used for an evaluation of small-scale diffusion experiments.

  13. Combining computational models, semantic annotations and simulation experiments in a graph database

    PubMed Central

    Henkel, Ron; Wolkenhauer, Olaf; Waltemath, Dagmar

    2015-01-01

    Model repositories such as the BioModels Database, the CellML Model Repository or JWS Online are frequently accessed to retrieve computational models of biological systems. However, their storage concepts support only restricted types of queries and not all data inside the repositories can be retrieved. In this article we present a storage concept that meets this challenge. It grounds on a graph database, reflects the models’ structure, incorporates semantic annotations and simulation descriptions and ultimately connects different types of model-related data. The connections between heterogeneous model-related data and bio-ontologies enable efficient search via biological facts and grant access to new model features. The introduced concept notably improves the access of computational models and associated simulations in a model repository. This has positive effects on tasks such as model search, retrieval, ranking, matching and filtering. Furthermore, our work for the first time enables CellML- and Systems Biology Markup Language-encoded models to be effectively maintained in one database. We show how these models can be linked via annotations and queried. Database URL: https://sems.uni-rostock.de/projects/masymos/ PMID:25754863

  14. High-level nuclear waste transport and storage assessment of potential impacts on tourism in the Las Vegas area. Nevada Nuclear Waste Storage Investigations Project

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    The literature review and empirical analyses presented in this report were undertaken, for the most part, between August and October 1983. They are not comprehensive. No primary data were gathered, nor were any formal surveys conducted. Additionally, because construction of a repository at Yucca Mountain, if that site is selected for a repository, is not scheduled to begin until 1993, engineering design and planned physical appearance of the repository are very preliminary. Therefore, specific design features or visual appearance were not addressed in the analyses. Finally, because actual transportation routes have not been designated, impacts on tourism generated specifically bymore » transportation activities are not considered separately. Chapter 2 briefly discusses possible means by which a repository could impact tourism in the Las Vegas area. Chapter 3 presents a review of previous research on alternative methods for predicting the response of people to potential hazards. A review of several published studies where these methods have been applied to facilities and activities associated with radioactive materials is included in Chapter 3. Chapter 4 contains five case studies of tourism impacts associated with past events that were perceived by the public to represent safety hazards. These perceptions of safety hazards were evidenced by news media coverage. These case studies were conducted specifically for this report. Conclusions of this preliminary analysis regarding the potential impact on tourism in the Las Vegas area of a repository at Yucca Mountain are in Chapter 5. Recommendations for further research are contained in Chapter 6.« less

  15. Proceedings of the scientific visit on crystalline rock repository development.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mariner, Paul E.; Hardin, Ernest L.; Miksova, Jitka

    2013-02-01

    A scientific visit on Crystalline Rock Repository Development was held in the Czech Republic on September 24-27, 2012. The visit was hosted by the Czech Radioactive Waste Repository Authority (RAWRA), co-hosted by Sandia National Laboratories (SNL), and supported by the International Atomic Energy Agency (IAEA). The purpose of the visit was to promote technical information exchange between participants from countries engaged in the investigation and exploration of crystalline rock for the eventual construction of nuclear waste repositories. The visit was designed especially for participants of countries that have recently commenced (or recommenced) national repository programmes in crystalline host rock formations.more » Discussion topics included repository programme development, site screening and selection, site characterization, disposal concepts in crystalline host rock, regulatory frameworks, and safety assessment methodology. Interest was surveyed in establishing a %E2%80%9Cclub,%E2%80%9D the mission of which would be to identify and address the various technical challenges that confront the disposal of radioactive waste in crystalline rock environments. The idea of a second scientific visit to be held one year later in another host country received popular support. The visit concluded with a trip to the countryside south of Prague where participants were treated to a tour of the laboratory and underground facilities of the Josef Regional Underground Research Centre.« less

  16. Research on Geo-information Data Model for Preselected Areas of Geological Disposal of High-level Radioactive Waste

    NASA Astrophysics Data System (ADS)

    Gao, M.; Huang, S. T.; Wang, P.; Zhao, Y. A.; Wang, H. B.

    2016-11-01

    The geological disposal of high-level radioactive waste (hereinafter referred to "geological disposal") is a long-term, complex, and systematic scientific project, whose data and information resources in the research and development ((hereinafter referred to ”R&D”) process provide the significant support for R&D of geological disposal system, and lay a foundation for the long-term stability and safety assessment of repository site. However, the data related to the research and engineering in the sitting of the geological disposal repositories is more complicated (including multi-source, multi-dimension and changeable), the requirements for the data accuracy and comprehensive application has become much higher than before, which lead to the fact that the data model design of geo-information database for the disposal repository are facing more serious challenges. In the essay, data resources of the pre-selected areas of the repository has been comprehensive controlled and systematic analyzed. According to deeply understanding of the application requirements, the research work has made a solution for the key technical problems including reasonable classification system of multi-source data entity, complex logic relations and effective physical storage structures. The new solution has broken through data classification and conventional spatial data the organization model applied in the traditional industry, realized the data organization and integration with the unit of data entities and spatial relationship, which were independent, holonomic and with application significant features in HLW geological disposal. The reasonable, feasible and flexible data conceptual models, logical models and physical models have been established so as to ensure the effective integration and facilitate application development of multi-source data in pre-selected areas for geological disposal.

  17. Shaping Solutions from Learnings in PAIs: A Blueprint

    ERIC Educational Resources Information Center

    Dosanjh, Nawtej; Jha, Pushkar P.

    2016-01-01

    Purpose: The paper outlines a portal that facilitates learning through sharing of experiences. This flow is between experience sharers and solution seekers in the domain of poverty alleviation interventions (PAIs). Practitioners working on PAIs are often confined to searching from within "lessons learned" repositories and also from…

  18. Mining databases for protein aggregation: a review.

    PubMed

    Tsiolaki, Paraskevi L; Nastou, Katerina C; Hamodrakas, Stavros J; Iconomidou, Vassiliki A

    2017-09-01

    Protein aggregation is an active area of research in recent decades, since it is the most common and troubling indication of protein instability. Understanding the mechanisms governing protein aggregation and amyloidogenesis is a key component to the aetiology and pathogenesis of many devastating disorders, including Alzheimer's disease or type 2 diabetes. Protein aggregation data are currently found "scattered" in an increasing number of repositories, since advances in computational biology greatly influence this field of research. This review exploits the various resources of aggregation data and attempts to distinguish and analyze the biological knowledge they contain, by introducing protein-based, fragment-based and disease-based repositories, related to aggregation. In order to gain a broad overview of the available repositories, a novel comprehensive network maps and visualizes the current association between aggregation databases and other important databases and/or tools and discusses the beneficial role of community annotation. The need for unification of aggregation databases in a common platform is also addressed.

  19. Provenance Storage, Querying, and Visualization in PBase

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kianmajd, Parisa; Ludascher, Bertram; Missier, Paolo

    2015-01-01

    We present PBase, a repository for scientific workflows and their corresponding provenance information that facilitates the sharing of experiments among the scientific community. PBase is interoperable since it uses ProvONE, a standard provenance model for scientific workflows. Workflows and traces are stored in RDF, and with the support of SPARQL and the tree cover encoding, the repository provides a scalable infrastructure for querying the provenance data. Furthermore, through its user interface, it is possible to: visualize workflows and execution traces; visualize reachability relations within these traces; issue SPARQL queries; and visualize query results.

  20. A Digital Broadcast Item (DBI) enabling metadata repository for digital, interactive television (digiTV) feedback channel networks

    NASA Astrophysics Data System (ADS)

    Lugmayr, Artur R.; Mailaparampil, Anurag; Tico, Florina; Kalli, Seppo; Creutzburg, Reiner

    2003-01-01

    Digital television (digiTV) is an additional multimedia environment, where metadata is one key element for the description of arbitrary content. This implies adequate structures for content description, which is provided by XML metadata schemes (e.g. MPEG-7, MPEG-21). Content and metadata management is the task of a multimedia repository, from which digiTV clients - equipped with an Internet connection - can access rich additional multimedia types over an "All-HTTP" protocol layer. Within this research work, we focus on conceptual design issues of a metadata repository for the storage of metadata, accessible from the feedback channel of a local set-top box. Our concept describes the whole heterogeneous life-cycle chain of XML metadata from the service provider to the digiTV equipment, device independent representation of content, accessing and querying the metadata repository, management of metadata related to digiTV, and interconnection of basic system components (http front-end, relational database system, and servlet container). We present our conceptual test configuration of a metadata repository that is aimed at a real-world deployment, done within the scope of the future interaction (fiTV) project at the Digital Media Institute (DMI) Tampere (www.futureinteraction.tv).

  1. Childhood vesicoureteral reflux studies: registries and repositories sources and nosology.

    PubMed

    Chesney, Russell W; Patters, Andrea B

    2013-12-01

    Despite several recent studies, the advisability of antimicrobial prophylaxis and certain imaging studies for urinary tract infections (UTIs) remains controversial. The role of vesicoureteral reflux (VUR) on the severity and re-infection rates for UTIs is also difficult to assess. Registries and repositories of data and biomaterials from clinical studies in children with VUR are valuable. Disease registries are collections of secondary data related to patients with a specific diagnosis, condition or procedure. Registries differ from indices in that they contain more extensive data. A research repository is an entity that receives, stores, processes and/or disseminates specimens (or other materials) as needed. It encompasses the physical location as well as the full range of activities associated with its operation. It may also be referred to as a biorepository. This report provides information about some current registries and repositories that include data and samples from children with VUR. It also describes the heterogeneous nature of the subjects, as some registries and repositories include only data or samples from patients with primary reflux while others also include those from patients with syndromic or secondary reflux. Copyright © 2012 Journal of Pediatric Urology Company. All rights reserved.

  2. Digital Rocks Portal: Preservation, Sharing, Remote Visualization and Automated Analysis of Imaged Datasets

    NASA Astrophysics Data System (ADS)

    Prodanovic, M.; Esteva, M.; Ketcham, R. A.; Hanlon, M.; Pettengill, M.; Ranganath, A.; Venkatesh, A.

    2016-12-01

    Due to advances in imaging modalities such as X-ray microtomography and scattered electron microscopy, 2D and 3D imaged datasets of rock microstructure on nanometer to centimeter length scale allow investigation of nonlinear flow and mechanical phenomena using numerical approaches. This in turn produces various upscaled parameters required by subsurface flow and deformation simulators. However, a single research group typically specializes in an imaging modality and/or related modeling on a single length scale, and lack of data-sharing infrastructure makes it difficult to integrate different length scales. We developed a sustainable, open and easy-to-use repository called the Digital Rocks Portal (http://www.digitalrocksportal.org), that (1) organizes images and related experimental measurements of different porous materials, (2) improves access to them for a wider community of geosciences or engineering researchers not necessarily trained in computer science or data analysis. Our objective is to enable scientific inquiry and engineering decisions founded on a data-driven basis. We show how the data loaded in the portal can be documented, referenced in publications via digital object identifiers, visualize and linked to other repositories. We then show preliminary results on integrating remote parallel visualization and flow simulation workflow with the pore structures currently stored in the repository. We finally discuss the issues of collecting correct metadata, data discoverability and repository sustainability. This is the first repository for this particular data, but is part of the wider ecosystem of geoscience data and model cyber-infrastructure called "Earthcube" (http://earthcube.org/) sponsored by National Science Foundation. For data sustainability and continuous access, the portal is implemented within the reliable, 24/7 maintained High Performance Computing Infrastructure supported by the Texas Advanced Computing Center (TACC) at the University of Texas at Austin. Long-term storage is provided through the University of Texas System Research Cyber-infrastructure initiative.

  3. The impact of the privacy rule on cancer research: variations in attitudes and application of regulatory standards.

    PubMed

    Goss, Elizabeth; Link, Michael P; Bruinooge, Suanna S; Lawrence, Theodore S; Tepper, Joel E; Runowicz, Carolyn D; Schilsky, Richard L

    2009-08-20

    The American Society of Clinical Oncology (ASCO) Cancer Research Committee designed a qualitative research project to assess the attitudes of cancer researchers and compliance officials regarding compliance with the US Privacy Rule and to identify potential strategies for eliminating perceived or real barriers to achieving compliance. A team of three interviewers asked 27 individuals (13 investigators and 14 compliance officials) from 13 institutions to describe the anticipated approach of their institutions to Privacy Rule compliance in three hypothetical research studies. The interviews revealed that although researchers and compliance officials share the view that patients' cancer diagnoses should enjoy a high level of privacy protection, there are significant tensions between the two groups related to the proper standards for compliance necessary to protect patients. The disagreements are seen most clearly with regard to the appropriate definition of a "future research use" of protected health information in biospecimen and data repositories and the standards for a waiver of authorization for disclosure and use of such data. ASCO believes that disagreements related to compliance and the resulting delays in certain projects and abandonment of others might be eased by additional institutional training programs and consultation on Privacy Rule issues during study design. ASCO also proposes the development of best practices documents to guide 1) creation of data repositories, 2) disclosure and use of data from such repositories, and 3) the design of survivorship and genetics studies.

  4. Making proteomics data accessible and reusable: Current state of proteomics databases and repositories

    PubMed Central

    Perez-Riverol, Yasset; Alpi, Emanuele; Wang, Rui; Hermjakob, Henning; Vizcaíno, Juan Antonio

    2015-01-01

    Compared to other data-intensive disciplines such as genomics, public deposition and storage of MS-based proteomics, data are still less developed due to, among other reasons, the inherent complexity of the data and the variety of data types and experimental workflows. In order to address this need, several public repositories for MS proteomics experiments have been developed, each with different purposes in mind. The most established resources are the Global Proteome Machine Database (GPMDB), PeptideAtlas, and the PRIDE database. Additionally, there are other useful (in many cases recently developed) resources such as ProteomicsDB, Mass Spectrometry Interactive Virtual Environment (MassIVE), Chorus, MaxQB, PeptideAtlas SRM Experiment Library (PASSEL), Model Organism Protein Expression Database (MOPED), and the Human Proteinpedia. In addition, the ProteomeXchange consortium has been recently developed to enable better integration of public repositories and the coordinated sharing of proteomics information, maximizing its benefit to the scientific community. Here, we will review each of the major proteomics resources independently and some tools that enable the integration, mining and reuse of the data. We will also discuss some of the major challenges and current pitfalls in the integration and sharing of the data. PMID:25158685

  5. BioAcoustica: a free and open repository and analysis platform for bioacoustics

    PubMed Central

    Baker, Edward; Price, Ben W.; Rycroft, S. D.; Smith, Vincent S.

    2015-01-01

    We describe an online open repository and analysis platform, BioAcoustica (http://bio.acousti.ca), for recordings of wildlife sounds. Recordings can be annotated using a crowdsourced approach, allowing voice introductions and sections with extraneous noise to be removed from analyses. This system is based on the Scratchpads virtual research environment, the BioVeL portal and the Taverna workflow management tool, which allows for analysis of recordings using a grid computing service. At present the analyses include spectrograms, oscillograms and dominant frequency analysis. Further analyses can be integrated to meet the needs of specific researchers or projects. Researchers can upload and annotate their recordings to supplement traditional publication. Database URL: http://bio.acousti.ca PMID:26055102

  6. Public involvement on closure of Asse II radioactive waste repository in Germany

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kallenbach-Herbert, Beate

    2013-07-01

    From 1967 to 1978, about 125,800 barrels of low- and intermediate level waste were disposed of - nominally for research purposes - in the former 'Asse' salt mine which had before been used for the production of potash for many years. Since 1988 an inflow of brine is being observed which will cause dangers of flooding and of a collapse due to salt weakening and dissolution if it should increase. Since several years the closure of the Asse repository is planned with the objective to prevent the flooding and collapse of the mine and the release of radioactive substances tomore » the biosphere. The first concept that was presented by the former operator, however, seemed completely unacceptable to regional representatives from politics and NGOs. Their activities against these plans made the project a top issue on the political agenda from the federal to the local level. The paper traces the main reasons which lead to the severe safety problems in the past as well as relevant changes in the governance system today. A focus is put on the process for public involvement in which the Citizens' Advisory Group 'A2B' forms the core measure. Its structure and framework, experience and results, expectations from inside and outside perspectives are presented. Furthermore the question is tackled how far this process can serve as an example for a participatory approach in a siting process for a geological repository for high active waste which can be expected to be highly contested in the affected regions. (authors)« less

  7. Health social workers sources of knowledge for decision making in practice.

    PubMed

    McDermott, Fiona; Henderson, Annabel; Quayle, Carol

    2017-10-01

    This article presents findings from research examining knowledge social workers in a health network in Victoria, Australia identified as informing their decision-making. Data for 13 patients, and in-depth interviews with six social workers who worked with these patients, were studied. A thematic analysis of interviews revealed that participants identified reliance on past experience and contextual/situational information as underpinning their decisions, demonstrating their commitment to person-in-environment perspectives. However, despite the availability of a repository of empirical evidence, no respondent made use of this. This study provided insight into health practitioners' sources of knowledge, highlighting gaps and areas for further exploration.

  8. Transcribing and digitizing eighteenth- and nineteenth-century letters for a historical digital repository.

    PubMed

    Dunster, Emily S; Kipnis, Daniel G; Angelo, F Michael

    2014-01-01

    In fall 2011, the Scott Memorial Library purchased 53 letters belonging to an 1841 graduate of Jefferson Medical College, John Plimpton Green. The library staff transcribed and digitized the letters, creating an online collection in the university's institutional repository, Jefferson Digital Commons. This article will detail the process of transcribing and digitizing the collection along with sharing statistics and the benefits of this project to global researchers.

  9. International Approaches for Nuclear Waste Disposal in Geological Formations: Report on Fifth Worldwide Review

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Faybishenko, Boris; Birkholzer, Jens; Persoff, Peter

    2016-09-01

    The goal of the Fifth Worldwide Review is to document evolution in the state-of-the-art of approaches for nuclear waste disposal in geological formations since the Fourth Worldwide Review that was released in 2006. The last ten years since the previous Worldwide Review has seen major developments in a number of nations throughout the world pursuing geological disposal programs, both in preparing and reviewing safety cases for the operational and long-term safety of proposed and operating repositories. The countries that are approaching implementation of geological disposal will increasingly focus on the feasibility of safely constructing and operating their repositories in short-more » and long terms on the basis existing regulations. The WWR-5 will also address a number of specific technical issues in safety case development along with the interplay among stakeholder concerns, technical feasibility, engineering design issues, and operational and post-closure safety. Preparation and publication of the Fifth Worldwide Review on nuclear waste disposal facilitates assessing the lessons learned and developing future cooperation between the countries. The Report provides scientific and technical experiences on preparing for and developing scientific and technical bases for nuclear waste disposal in deep geologic repositories in terms of requirements, societal expectations and the adequacy of cases for long-term repository safety. The Chapters include potential issues that may arise as repository programs mature, and identify techniques that demonstrate the safety cases and aid in promoting and gaining societal confidence. The report will also be used to exchange experience with other fields of industry and technology, in which concepts similar to the design and safety cases are applied, as well to facilitate the public perception and understanding of the safety of the disposal approaches relative to risks that may increase over long times frames in the absence of a successful implementation of final dispositioning.« less

  10. Preservation of Earth Science Data History with Digital Content Repository Technology

    NASA Astrophysics Data System (ADS)

    Wei, Y.; Pan, J.; Shrestha, B.; Cook, R. B.

    2011-12-01

    An increasing need for derived and on-demand data product in Earth Science research makes the digital content more difficult for providers to manage and preserve and for users to locate, understand, and consume. Specifically, this increasing need presents additional challenges in managing data processing history information and delivering such information to end users. For example, the North American Carbon Program (NACP) Multi-scale Synthesis and Terrestrial Model Intercomparison Project (MsTMIP) chose a modified SYNMAP land cover data as one of the input driver data for participating terrestrial biospheric models. The global 1km resolution SYNMAP data was created by harmonizing 3 remote sensing-based land cover products: GLCC, GLC2000, and the MODIS land cover product. The original SYNMAP land cover data was aggregated into half and quarter degree resolution. It was then enhanced with more detailed grassland and cropland types. Currently, there lacks an effective mechanism to convey this data processing information to different modeling teams for them to determine if a data product meets their needs. It still highly relies on offline human interaction. The NASA-sponsored ORNL DAAC has leveraged the contemporary digital object repository technology to promote the representation, management, and delivery of data processing history and provenance information. Within digital object repository, different data products are managed as objects, with metadata as attributes and content delivery and management services as dissemination methods. Derivation relationships among data products can be semantically referenced between digital objects. Within the repository, data users can easily track a derived data product back to its origin, explorer metadata and documents about each intermediate data product, and discover processing details involved in each derivation step. Coupled with Drupal Web Content Management System, the digital repository interface was enhanced to provide intuitive graphic representation of the data processing history. Each data product is also associated with a formal metadata record in FGDC standards, and the main fields of the FGDC record are indexed for search, and are displayed as attributes of the data product. These features enable data users to better understand and consume a data product. The representation of data processing history in digital repository can further promote long-term data preservation. Lineage information is a major aspect to make digital data understandable and usable long time into the future. Derivation references can be setup between digital objects not only within a single digital repository, but also across multiple distributed digital repositories. Along with emerging identification mechanisms, such as Digital Object Identifier (DOI), a flexible distributed digital repository network can be setup to better preserve digital content. In this presentation, we describe how digital content repository technology can be used to manage, preserve, and deliver digital data processing history information in Earth Science research domain, with selected data archived in ORNL DAAC and Model and Synthesis Thematic Data Center (MAST-DC) as testing targets.

  11. Geoengineering properties of potential repository units at Yucca Mountain, southern Nevada

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tillerson, J.R.; Nimick, F.B.

    1984-12-01

    The Nevada Nuclear Waste Storage Investigations (NNWSI) Project is currently evaluating volcanic tuffs at the Yucca Mountain site, located on and adjacent to the Nevada Test Site, for possible use as a host rock for a radioactive waste repository. The behavior of tuff as an engineering material must be understood to design, license, construct, and operate a repository. Geoengineering evaluations and measurements are being made to develop confidence in both the analysis techniques for thermal, mechanical, and hydrothermal effects and the supporting data base of rock properties. The analysis techniques and the data base are currently used for repository design,more » waste package design, and performance assessment analyses. This report documents the data base of geoengineering properties used in the analyses that aided the selection of the waste emplacement horizon and in analyses synopsized in the Environmental Assessment Report prepared for the Yucca Mountain site. The strategy used for the development of the data base relies primarily on data obtained in laboratory tests that are then confirmed in field tests. Average thermal and mechanical properties (and their anticipated variations) are presented. Based upon these data, analyses completed to date, and previous excavation experience in tuff, it is anticipated that existing mining technology can be used to develop stable underground openings and that repository operations can be carried out safely.« less

  12. Metadata Repository for Improved Data Sharing and Reuse Based on HL7 FHIR.

    PubMed

    Ulrich, Hannes; Kock, Ann-Kristin; Duhm-Harbeck, Petra; Habermann, Jens K; Ingenerf, Josef

    2016-01-01

    Unreconciled data structures and formats are a common obstacle to the urgently required sharing and reuse of data within healthcare and medical research. Within the North German Tumor Bank of Colorectal Cancer, clinical and sample data, based on a harmonized data set, is collected and can be pooled by using a hospital-integrated Research Data Management System supporting biobank and study management. Adding further partners who are not using the core data set requires manual adaptations and mapping of data elements. Facing this manual intervention and focusing the reuse of heterogeneous healthcare instance data (value level) and data elements (metadata level), a metadata repository has been developed. The metadata repository is an ISO 11179-3 conformant server application built for annotating and mediating data elements. The implemented architecture includes the translation of metadata information about data elements into the FHIR standard using the FHIR Data Element resource with the ISO 11179 Data Element Extensions. The FHIR-based processing allows exchange of data elements with clinical and research IT systems as well as with other metadata systems. With increasingly annotated and harmonized data elements, data quality and integration can be improved for successfully enabling data analytics and decision support.

  13. New Catalog of Resources Enables Paleogeosciences Research

    NASA Astrophysics Data System (ADS)

    Lingo, R. C.; Horlick, K. A.; Anderson, D. M.

    2014-12-01

    The 21st century promises a new era for scientists of all disciplines, the age where cyber infrastructure enables research and education and fuels discovery. EarthCube is a working community of over 2,500 scientists and students of many Earth Science disciplines who are looking to build bridges between disciplines. The EarthCube initiative will create a digital infrastructure that connects databases, software, and repositories. A catalog of resources (databases, software, repositories) has been produced by the Research Coordination Network for Paleogeosciences to improve the discoverability of resources. The Catalog is currently made available within the larger-scope CINERGI geosciences portal (http://hydro10.sdsc.edu/geoportal/catalog/main/home.page). Other distribution points and web services are planned, using linked data, content services for the web, and XML descriptions that can be harvested using metadata protocols. The databases provide searchable interfaces to find data sets that would otherwise remain dark data, hidden in drawers and on personal computers. The software will be described in catalog entries so just one click will lead users to methods and analytical tools that many geoscientists were unaware of. The repositories listed in the Paleogeosciences Catalog contain physical samples found all across the globe, from natural history museums to the basements of university buildings. EarthCube has over 250 databases, 300 software systems, and 200 repositories which will grow in the coming year. When completed, geoscientists across the world will be connected into a productive workflow for managing, sharing, and exploring geoscience data and information that expedites collaboration and innovation within the paleogeosciences, potentially bringing about new interdisciplinary discoveries.

  14. A Virtual Rock Physics Laboratory Through Visualized and Interactive Experiments

    NASA Astrophysics Data System (ADS)

    Vanorio, T.; Di Bonito, C.; Clark, A. C.

    2014-12-01

    As new scientific challenges demand more comprehensive and multidisciplinary investigations, laboratory experiments are not expected to become simpler and/or faster. Experimental investigation is an indispensable element of scientific inquiry and must play a central role in the way current and future generations of scientist make decisions. To turn the complexity of laboratory work (and that of rocks!) into dexterity, engagement, and expanded learning opportunities, we are building an interactive, virtual laboratory reproducing in form and function the Stanford Rock Physics Laboratory, at Stanford University. The objective is to combine lectures on laboratory techniques and an online repository of visualized experiments consisting of interactive, 3-D renderings of equipment used to measure properties central to the study of rock physics (e.g., how to saturate rocks, how to measure porosity, permeability, and elastic wave velocity). We use a game creation system together with 3-D computer graphics, and a narrative voice to guide the user through the different phases of the experimental protocol. The main advantage gained in employing computer graphics over video footage is that students can virtually open the instrument, single out its components, and assemble it. Most importantly, it helps describe the processes occurring within the rock. These latter cannot be tracked while simply recording the physical experiment, but computer animation can efficiently illustrate what happens inside rock samples (e.g., describing acoustic waves, and/or fluid flow through a porous rock under pressure within an opaque core-holder - Figure 1). The repository of visualized experiments will complement lectures on laboratory techniques and constitute an on-line course offered through the EdX platform at Stanford. This will provide a virtual laboratory for anyone, anywhere to facilitate teaching/learning of introductory laboratory classes in Geophysics and expand the number of courses that can be offered for curricula in Earth Sciences. The primary goal is to open up a research laboratory such as the one available at Stanford to promising students worldwide who are currently left out of such educational resources.

  15. LungMAP: The Molecular Atlas of Lung Development Program.

    PubMed

    Ardini-Poleske, Maryanne E; Clark, Robert F; Ansong, Charles; Carson, James P; Corley, Richard A; Deutsch, Gail H; Hagood, James S; Kaminski, Naftali; Mariani, Thomas J; Potter, Steven S; Pryhuber, Gloria S; Warburton, David; Whitsett, Jeffrey A; Palmer, Scott M; Ambalavanan, Namasivayam

    2017-11-01

    The National Heart, Lung, and Blood Institute is funding an effort to create a molecular atlas of the developing lung (LungMAP) to serve as a research resource and public education tool. The lung is a complex organ with lengthy development time driven by interactive gene networks and dynamic cross talk among multiple cell types to control and coordinate lineage specification, cell proliferation, differentiation, migration, morphogenesis, and injury repair. A better understanding of the processes that regulate lung development, particularly alveologenesis, will have a significant impact on survival rates for premature infants born with incomplete lung development and will facilitate lung injury repair and regeneration in adults. A consortium of four research centers, a data coordinating center, and a human tissue repository provides high-quality molecular data of developing human and mouse lungs. LungMAP includes mouse and human data for cross correlation of developmental processes across species. LungMAP is generating foundational data and analysis, creating a web portal for presentation of results and public sharing of data sets, establishing a repository of young human lung tissues obtained through organ donor organizations, and developing a comprehensive lung ontology that incorporates the latest findings of the consortium. The LungMAP website (www.lungmap.net) currently contains more than 6,000 high-resolution lung images and transcriptomic, proteomic, and lipidomic human and mouse data and provides scientific information to stimulate interest in research careers for young audiences. This paper presents a brief description of research conducted by the consortium, database, and portal development and upcoming features that will enhance the LungMAP experience for a community of users. Copyright © 2017 the American Physiological Society.

  16. Genomic research and wide data sharing: views of prospective participants.

    PubMed

    Trinidad, Susan Brown; Fullerton, Stephanie M; Bares, Julie M; Jarvik, Gail P; Larson, Eric B; Burke, Wylie

    2010-08-01

    Sharing study data within the research community generates tension between two important goods: promoting scientific goals and protecting the privacy interests of study participants. This study was designed to explore the perceptions, beliefs, and attitudes of research participants and possible future participants regarding genome-wide association studies and repository-based research. Focus group sessions with (1) current research participants, (2) surrogate decision-makers, and (3) three age-defined cohorts (18-34 years, 35-50, >50). Participants expressed a variety of opinions about the acceptability of wide sharing of genetic and phenotypic information for research purposes through large, publicly accessible data repositories. Most believed that making de-identified study data available to the research community is a social good that should be pursued. Privacy and confidentiality concerns were common, although they would not necessarily preclude participation. Many participants voiced reservations about sharing data with for-profit organizations. Trust is central in participants' views regarding data sharing. Further research is needed to develop governance models that enact the values of stewardship.

  17. The New Face of Data Accessibility

    NASA Technical Reports Server (NTRS)

    Fitts, Mary A.; VanBaalan, Mary; Johnson-Throop, Kathy A.; Thomas, Deidre; Havelka, Jacque

    2010-01-01

    Management of medical and research data at NASA's Johnson Space Center has been addressed with two separate, independent systems: the Lifetime Surveillance of Astronaut Health (formerly, The Longitudinal Study of Astronaut Health) (LSAH) and the Life Sciences Data Archive (LSDA). Project management for these has been autonomous with little or no cross-over of goals, objectives or strategy. The result has been limited debate and discussion regarding how contents from one repository might impact or guide the direction of the other. It is decidedly more efficient to use existing data and information than to re-generate them. Ensuring that both clinical and research data / information are accessible for review is a central concept to the decision to unify these repositories. In the past, research data from flight and ground analogs has been held in the LSDA and medical data held in the Electronic Medical Record or in console flight surgeon logs and records. There was little cross-pollination between medical and research findings and, as a result, applicable research was not being fully incorporated into clinical, in-flight practice. Conversely, findings by the console surgeon were not being picked up by the research community. The desired life cycle for risk mitigation was not being fully realized. The goal of unifying these repositories and processes is to provide a closely knit approach to handling medical and research data, which will not only engender discussion and debate but will also ensure that both categories of data and information are used to enhance the use of medical and research data to reduce risk and promote the understanding of space physiology, countermeasures and other mitigation strategies

  18. Nature Research journals reproducibility policies and initiatives in the Earth sciences

    NASA Astrophysics Data System (ADS)

    VanDecar, J. C.

    2016-12-01

    The Nature Research journals strongly support the long-term endeavour by funders, institutions, researchers and publishers toward increasing the reliability and reproducibility of published research. In the Earth, space and environmental sciences this mainly takes the form of ensuring that underlying data and methods in each manuscript are made as transparent and accessible as possible. Supporting data must be made available to editors and peer reviewers at the time of submission for the purposes of evaluating each manuscript. But the preferred way to share data sets is via public repositories. When appropriate community repositories are available, we strongly encourage authors to deposit their data prior to publication. We also now require that a statement be included in each manuscript, under the heading "Data availability", indicating whether and how the data can be accessed, including any restrictions to access. To allow authors to describe their experimental design and methods in as much detail as necessary, the Nature Research journals have effectively abolished space restrictions on online methods sections. To further increase transparency, we also encourage authors to provide tables of the data behind graphs and figures as Source Data. This builds on our established data-deposition policy for specific experiments and large data sets. The Source Data is made available directly from the figure legend, for easy access. We also require that details of geological samples and palaeontological specimens include clear provenance information to ensure full transparency of the research methods. Palaeontological and type specimens must be deposited in a recognised museum or collection to permit free access by other researchers in perpetuity. Finally, authors must make available upon request, to editors and reviewers, any previously unreported custom computer code used to generate results that are reported in the paper and central to its main claims. For all studies using custom code that is deemed central to the conclusions, a statement must be included, under the heading "Code availability", indicating whether and how the code can be accessed, including any restrictions to access.

  19. National Programs | Frederick National Laboratory for Cancer Research

    Cancer.gov

    The Frederick National Laboratoryis a shared national resource that offers access to a suite of advanced biomedical technologies, provides selected science and technology services, and maintains vast repositories of research materials available

  20. Eagle-i: Making Invisible Resources, Visible

    PubMed Central

    Haendel, M.; Wilson, M.; Torniai, C.; Segerdell, E.; Shaffer, C.; Frost, R.; Bourges, D.; Brownstein, J.; McInnerney, K.

    2010-01-01

    RP-134 The eagle-i Consortium – Dartmouth College, Harvard Medical School, Jackson State University, Morehouse School of Medicine, Montana State University, Oregon Health and Science University (OHSU), the University of Alaska, the University of Hawaii, and the University of Puerto Rico – aims to make invisible resources for scientific research visible by developing a searchable network of resource repositories at research institutions nationwide. Now in early development, it is hoped that the system will scale beyond the consortium at the end of the two-year pilot. Data Model & Ontology: The eagle-i ontology development team at the OHSU Library is generating the data model and ontologies necessary for resource indexing and querying. Our indexing system will enable cores and research labs to represent resources within a defined vocabulary, leading to more effective searches and better linkage between data types. This effort is being guided by active discussions within the ontology community (http://RRontology.tk) bringing together relevant preexisting ontologies in a logical framework. The goal of these discussions is to provide context for interoperability and domain-wide standards for resource types used throughout biomedical research. Research community feedback is welcomed. Architecture Development, led by a team at Harvard, includes four main components: tools for data collection, management and curation; an institutional resource repository; a federated network; and a central search application. Each participating institution will populate and manage their repository locally, using data collection and curation tools. To help improve search performance, data tools will support the semi-automatic annotation of resources. A central search application will use a federated protocol to broadcast queries to all repositories and display aggregated results. The search application will leverage the eagle-i ontologies to help guide users to valid queries via auto-suggestions and taxonomy browsing and improve search result quality via concept-based search and synonym expansion. Website: http://eagle-i.org. NIH/NCRR ARRA award #U24RR029825

  1. United States National Sewage Sludge Repository at Arizona State University--a new resource and research tool for environmental scientists, engineers, and epidemiologists.

    PubMed

    Venkatesan, Arjun K; Done, Hansa Y; Halden, Rolf U

    2015-02-01

    Processed municipal sewage sludges (MSS) are an abundant, unwanted by-product of wastewater treatment, increasingly applied to agriculture and forestry for inexpensive disposal and soil conditioning. Due to their high organic carbon and lipid contents, MSS not only is rich in carbon and nutrients but also represents a "sink" for recalcitrant, hydrophobic, and potentially bioaccumulative compounds. Indeed, many organics sequestered and concentrated in MSS meet the US Environmental Protection Agency's definition of being persistent, bioaccumulative, and toxic (PBT). In a strategic effort, our research team at the Biodesign Institute has created the National Sewage Sludge Repository (NSSR), a large repository of digested MSSs from 164 wastewater treatment plants from across the USA, as part of the Human Health Observatory (H2O) at Arizona State University (ASU). The NSSR likely represents the largest archive of digested MSS specimens in the USA. The present study summarizes key findings gleaned thus far from analysis of NSSR samples. For example, we evaluated the content of toxicants in MSS and computed estimates of nationwide inventories of mass produced chemicals that become sequestrated in sludge and later are released into the environment during sludge disposal on land. Ongoing efforts document co-occurrence of a variety of PBT compounds in both MSS and human samples, while also identifying a large number of potentially harmful MSS constituents for which human exposure data are still lacking. Finally, we summarize future opportunities and invite collaborative use of the NSSR by the research community. The H2O at ASU represents a new resource and research tool for environmental scientists and the larger research community. As illustrated in this work, this repository can serve to (i) identify and prioritize emerging contaminants, (ii) provide spatial and temporal trends of contaminants, (iii) inform and evaluate the effectiveness of environmental policy-making and regulations, and (iv) approximate, ongoing exposures and body burdens of mass-produced chemicals in human society.

  2. United States National Sewage Sludge Repository at Arizona State University – A New Resource and Research Tool for Environmental Scientists, Engineers, and Epidemiologists

    PubMed Central

    Venkatesan, Arjun K.; Done, Hansa Y.; Halden, Rolf U.

    2014-01-01

    Processed municipal sewage sludges (MSS) are an abundant, unwanted by-product of wastewater treatment, increasingly applied to agriculture and forestry for inexpensive disposal and soil conditioning. Due to their high organic-carbon and lipid contents, MSS not only is rich in carbon and nutrients but also represents a ‘sink’ for recalcitrant, hydrophobic and potentially bioaccumulative compounds. Indeed, many organics sequestered and concentrated in MSS meet the U.S. Environmental Protection Agency's definition of being persistent, bioaccumulative, and toxic (PBT). In a strategic effort, our research team at the Biodesign Institute has created the National Sewage Sludge Repository (NSSR), a large repository of digested MSSs from 164 wastewater treatment plants from across the USA, as part of the Human Health Observatory (H2O) at Arizona State University (ASU). The NSSR likely represents the largest archive of digested MSS specimens in the USA. The present study summarizes key findings gleaned thus far from analysis of NSSR samples. For example, we evaluated the content of toxicants in MSS and computed estimates of nationwide inventories of mass produced chemicals that become sequestrated in sludge and later are released into the environment during sludge disposal on land. Ongoing efforts document co-occurrence of a variety of PBT compounds in both MSS and human samples, while also identifying a large number of potentially harmful MSS constituents for which human exposure data are still lacking. Finally, we summarize new future opportunities and invite collaborative use the NSSR by the research community. The H2O at ASU represents a resource and research tool for environmental scientists and the larger research community. As illustrated in this work, this repository can serve to (i) identify and prioritize emerging contaminants; (ii) provide spatial and temporal trends of contaminants; (iii) inform and evaluate the effectiveness of environmental policy-making and regulations; and (iv) approximate, ongoing exposures and body burdens of mass-produced chemicals in human society. PMID:24824503

  3. Assessment of the long-term durability of concrete in radioactive waste repositories

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Atkinson, A.; Goult, D.J.; Hearne, J.A.

    1986-01-01

    A preliminary assessment of the long-term durability of concrete in a repository sited in clay is presented. The assessment is based on recorded experience of concrete structures and both field and laboratory studies. It is also supported by results of the examination of a concrete sample which had been buried in clay for 43 years. The engineering lifetime of a 1 m thick reinforced concrete slab, with one face in contact with clay, and the way in which pH in the repository as a whole is likely to vary with time have both been estimated from available data. The estimatesmore » indicate that engineering lifetimes of about 10/sup 3/ years are expected (providing that sulfate resisting cement is used) and that pH is likely to remain above 10.5 for about 10/sup 6/ years.« less

  4. Experimental and numerical simulation of dissolution and precipitation: implications for fracture sealing at Yucca Mountain, Nevada

    NASA Astrophysics Data System (ADS)

    Dobson, Patrick F.; Kneafsey, Timothy J.; Sonnenthal, Eric L.; Spycher, Nicolas; Apps, John A.

    2003-05-01

    Plugging of flow paths caused by mineral precipitation in fractures above the potential repository at Yucca Mountain, Nevada could reduce the probability of water seeping into the repository. As part of an ongoing effort to evaluate thermal-hydrological-chemical (THC) effects on flow in fractured media, we performed a laboratory experiment and numerical simulations to investigate mineral dissolution and precipitation under anticipated temperature and pressure conditions in the repository. To replicate mineral dissolution by vapor condensate in fractured tuff, water was flowed through crushed Yucca Mountain tuff at 94 °C. The resulting steady-state fluid composition had a total dissolved solids content of about 140 mg/l; silica was the dominant dissolved constituent. A portion of the steady-state mineralized water was flowed into a vertically oriented planar fracture in a block of welded Topopah Spring Tuff that was maintained at 80 °C at the top and 130 °C at the bottom. The fracture began to seal with amorphous silica within 5 days. A 1-D plug-flow numerical model was used to simulate mineral dissolution, and a similar model was developed to simulate the flow of mineralized water through a planar fracture, where boiling conditions led to mineral precipitation. Predicted concentrations of the major dissolved constituents for the tuff dissolution were within a factor of 2 of the measured average steady-state compositions. The mineral precipitation simulations predicted the precipitation of amorphous silica at the base of the boiling front, leading to a greater than 50-fold decrease in fracture permeability in 5 days, consistent with the laboratory experiment. These results help validate the use of a numerical model to simulate THC processes at Yucca Mountain. The experiment and simulations indicated that boiling and concomitant precipitation of amorphous silica could cause significant reductions in fracture porosity and permeability on a local scale. However, differences in fluid flow rates and thermal gradients between the experimental setup and anticipated conditions at Yucca Mountain need to be factored into scaling the results of the dissolution/precipitation experiments and associated simulations to THC models for the potential Yucca Mountain repository.

  5. OWLing Clinical Data Repositories With the Ontology Web Language

    PubMed Central

    Pastor, Xavier; Lozano, Esther

    2014-01-01

    Background The health sciences are based upon information. Clinical information is usually stored and managed by physicians with precarious tools, such as spreadsheets. The biomedical domain is more complex than other domains that have adopted information and communication technologies as pervasive business tools. Moreover, medicine continuously changes its corpus of knowledge because of new discoveries and the rearrangements in the relationships among concepts. This scenario makes it especially difficult to offer good tools to answer the professional needs of researchers and constitutes a barrier that needs innovation to discover useful solutions. Objective The objective was to design and implement a framework for the development of clinical data repositories, capable of facing the continuous change in the biomedicine domain and minimizing the technical knowledge required from final users. Methods We combined knowledge management tools and methodologies with relational technology. We present an ontology-based approach that is flexible and efficient for dealing with complexity and change, integrated with a solid relational storage and a Web graphical user interface. Results Onto Clinical Research Forms (OntoCRF) is a framework for the definition, modeling, and instantiation of data repositories. It does not need any database design or programming. All required information to define a new project is explicitly stated in ontologies. Moreover, the user interface is built automatically on the fly as Web pages, whereas data are stored in a generic repository. This allows for immediate deployment and population of the database as well as instant online availability of any modification. Conclusions OntoCRF is a complete framework to build data repositories with a solid relational storage. Driven by ontologies, OntoCRF is more flexible and efficient to deal with complexity and change than traditional systems and does not require very skilled technical people facilitating the engineering of clinical software systems. PMID:25599697

  6. OWLing Clinical Data Repositories With the Ontology Web Language.

    PubMed

    Lozano-Rubí, Raimundo; Pastor, Xavier; Lozano, Esther

    2014-08-01

    The health sciences are based upon information. Clinical information is usually stored and managed by physicians with precarious tools, such as spreadsheets. The biomedical domain is more complex than other domains that have adopted information and communication technologies as pervasive business tools. Moreover, medicine continuously changes its corpus of knowledge because of new discoveries and the rearrangements in the relationships among concepts. This scenario makes it especially difficult to offer good tools to answer the professional needs of researchers and constitutes a barrier that needs innovation to discover useful solutions. The objective was to design and implement a framework for the development of clinical data repositories, capable of facing the continuous change in the biomedicine domain and minimizing the technical knowledge required from final users. We combined knowledge management tools and methodologies with relational technology. We present an ontology-based approach that is flexible and efficient for dealing with complexity and change, integrated with a solid relational storage and a Web graphical user interface. Onto Clinical Research Forms (OntoCRF) is a framework for the definition, modeling, and instantiation of data repositories. It does not need any database design or programming. All required information to define a new project is explicitly stated in ontologies. Moreover, the user interface is built automatically on the fly as Web pages, whereas data are stored in a generic repository. This allows for immediate deployment and population of the database as well as instant online availability of any modification. OntoCRF is a complete framework to build data repositories with a solid relational storage. Driven by ontologies, OntoCRF is more flexible and efficient to deal with complexity and change than traditional systems and does not require very skilled technical people facilitating the engineering of clinical software systems.

  7. FitSearch: a robust way to interpret a yeast fitness profile in terms of drug's mode-of-action.

    PubMed

    Lee, Minho; Han, Sangjo; Chang, Hyeshik; Kwak, Youn-Sig; Weller, David M; Kim, Dongsup

    2013-01-01

    Yeast deletion-mutant collections have been successfully used to infer the mode-of-action of drugs especially by profiling chemical-genetic and genetic-genetic interactions on a genome-wide scale. Although tens of thousands of those profiles are publicly available, a lack of an accurate method for mining such data has been a major bottleneck for more widespread use of these useful resources. For general usage of those public resources, we designed FitRankDB as a general repository of fitness profiles, and developed a new search algorithm, FitSearch, for identifying the profiles that have a high similarity score with statistical significance for a given fitness profile. We demonstrated that our new repository and algorithm are highly beneficial to researchers who attempting to make hypotheses based on unknown modes-of-action of bioactive compounds, regardless of the types of experiments that have been performed using yeast deletion-mutant collection in various types of different measurement platforms, especially non-chip-based platforms. We showed that our new database and algorithm are useful when attempting to construct a hypothesis regarding the unknown function of a bioactive compound through small-scale experiments with a yeast deletion collection in a platform independent manner. The FitRankDB and FitSearch enhance the ease of searching public yeast fitness profiles and obtaining insights into unknown mechanisms of action of drugs. FitSearch is freely available at http://fitsearch.kaist.ac.kr.

  8. FitSearch: a robust way to interpret a yeast fitness profile in terms of drug's mode-of-action

    PubMed Central

    2013-01-01

    Background Yeast deletion-mutant collections have been successfully used to infer the mode-of-action of drugs especially by profiling chemical-genetic and genetic-genetic interactions on a genome-wide scale. Although tens of thousands of those profiles are publicly available, a lack of an accurate method for mining such data has been a major bottleneck for more widespread use of these useful resources. Results For general usage of those public resources, we designed FitRankDB as a general repository of fitness profiles, and developed a new search algorithm, FitSearch, for identifying the profiles that have a high similarity score with statistical significance for a given fitness profile. We demonstrated that our new repository and algorithm are highly beneficial to researchers who attempting to make hypotheses based on unknown modes-of-action of bioactive compounds, regardless of the types of experiments that have been performed using yeast deletion-mutant collection in various types of different measurement platforms, especially non-chip-based platforms. Conclusions We showed that our new database and algorithm are useful when attempting to construct a hypothesis regarding the unknown function of a bioactive compound through small-scale experiments with a yeast deletion collection in a platform independent manner. The FitRankDB and FitSearch enhance the ease of searching public yeast fitness profiles and obtaining insights into unknown mechanisms of action of drugs. FitSearch is freely available at http://fitsearch.kaist.ac.kr. PMID:23368702

  9. Improving global data infrastructures for more effective and scalable analysis of Earth and environmental data: the Australian NCI NERDIP Approach

    NASA Astrophysics Data System (ADS)

    Evans, Ben; Wyborn, Lesley; Druken, Kelsey; Richards, Clare; Trenham, Claire; Wang, Jingbo; Rozas Larraondo, Pablo; Steer, Adam; Smillie, Jon

    2017-04-01

    The National Computational Infrastructure (NCI) facility hosts one of Australia's largest repositories (10+ PBytes) of research data collections spanning datasets from climate, coasts, oceans, and geophysics through to astronomy, bioinformatics, and the social sciences domains. The data are obtained from national and international sources, spanning a wide range of gridded and ungridded (i.e., line surveys, point clouds) data, and raster imagery, as well as diverse coordinate reference projections and resolutions. Rather than managing these data assets as a digital library, whereby users can discover and download files to personal servers (similar to borrowing 'books' from a 'library'), NCI has built an extensive and well-integrated research data platform, the National Environmental Research Data Interoperability Platform (NERDIP, http://nci.org.au/data-collections/nerdip/). The NERDIP architecture enables programmatic access to data via standards-compliant services for high performance data analysis, and provides a flexible cloud-based environment to facilitate the next generation of transdisciplinary scientific research across all data domains. To improve use of modern scalable data infrastructures that are focused on efficient data analysis, the data organisation needs to be carefully managed including performance evaluations of projections and coordinate systems, data encoding standards and formats. A complication is that we have often found multiple domain vocabularies and ontologies are associated with equivalent datasets. It is not practical for individual dataset managers to determine which standards are best to apply to their dataset as this could impact accessibility and interoperability. Instead, they need to work with data custodians across interrelated communities and, in partnership with the data repository, the international scientific community to determine the most useful approach. For the data repository, this approach is essential to enable different disciplines and research communities to invoke new forms of analysis and discovery in an increasingly complex data-rich environment. Driven by the heterogeneity of Earth and environmental datasets, NCI developed a Data Quality/Data Assurance Strategy to ensure consistency is maintained within and across all datasets, as well as functionality testing to ensure smooth interoperability between products, tools, and services. This is particularly so for collections that contain data generated from multiple data acquisition campaigns, often using instruments and models that have evolved over time. By implementing the NCI Data Quality Strategy we have seen progressive improvement in the integration and quality of the datasets across the different subject domains, and through this, the ease by which the users can access data from this major data infrastructure. By both adhering to international standards and also contributing to extensions of these standards, data from the NCI NERDIP platform can be federated with data from other globally distributed data repositories and infrastructures. The NCI approach builds on our experience working with the astronomy and climate science communities, which have been internationally coordinating such interoperability standards within their disciplines for some years. The results of our work so far demonstrate more could be done in the Earth science, solid earth and environmental communities, particularly through establishing better linkages between international/national community efforts such as EPOS, ENVRIplus, EarthCube, AuScope and the Research Data Alliance.

  10. Proceedings of the 8th US/German Workshop on Salt Repository Research Design and Operation.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hansen, Francis D.; Steininger, Walter; Bollingerfehr, Wilhelm

    This document records the Proceedings of the 2017 gathering of salt repository nations. In a spirit of mutual support, technical issues are dissected, led capably by subject matter experts. As before, it is not possible to explore all contemporary issues regarding nuclear waste disposal in salt formations. Instead, the group focused on a few selected issues to be pursued in depth, while at the same time acknowledging and recording ancillary issues.

  11. Index of Oral Histories Relating to Naval Research and Development

    DTIC Science & Technology

    1985-01-01

    Repositories: NWC, DTNSRDC, NHC Individuals mentioned: Amlie, Dr. Thomas S. LaBerge , Dr. Walter McLean, Dr. William B. Parsons, RADM William S. Smith...future of R&D in the Navy. Repositories: NWC, DTNSRDC, NHC Individuals mentioned: Bennett, Dr. Ira Hollingsworth, Dr. Guilford L. LaBerge , Dr. Walter...DTNSRDC, NHC Individuals mentioned: Hunter, Dr. Hugh LaBerge . Dr. Walter McLean, Dr. William B. Brode, Dr. Wallace C. Sage, Dr. Bruce Wilson, Dr. Haskell

  12. Implementation of the Brazilian National Repository - RBMN Project - 13008

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Cassia Oliveira de Tello, Cledola

    2013-07-01

    Ionizing radiation in Brazil is used in electricity generation, medicine, industry, agriculture and for research and development purposes. All these activities can generate radioactive waste. At this point, in Brazil, the use of nuclear energy and radioisotopes justifies the construction of a national repository for radioactive wastes of low and intermediate-level. According to Federal Law No. 10308, Brazilian National Commission for Nuclear Energy (CNEN) is responsible for designing and constructing the intermediate and final storages for radioactive wastes. Additionally, a restriction on the construction of Angra 3 is that the repository is under construction until its operation start, attaining somemore » requirements of the Brazilian Environmental Regulator (IBAMA). Besides this NPP, in the National Energy Program is previewed the installation of four more plants, by 2030. In November 2008, CNEN launched the Project RBMN (Repository for Low and Intermediate-Level Radioactive Wastes), which aims at the implantation of a National Repository for disposal of low and intermediate-level of radiation wastes. This Project has some aspects that are unique in the Brazilian context, especially referring to the time between its construction and the end of its institutional period. This time is about 360 years, when the area will be released for unrestricted uses. It means that the Repository must be safe and secure for more than three hundred years, which is longer than half of the whole of Brazilian history. This aspect is very new for the Brazilian people, bringing a new dimension to public acceptance. Another point is this will be the first repository in South America, bringing a real challenge for the continent. The current status of the Project is summarized. (authors)« less

  13. Coupled Multi-physical Simulations for the Assessment of Nuclear Waste Repository Concepts: Modeling, Software Development and Simulation

    NASA Astrophysics Data System (ADS)

    Massmann, J.; Nagel, T.; Bilke, L.; Böttcher, N.; Heusermann, S.; Fischer, T.; Kumar, V.; Schäfers, A.; Shao, H.; Vogel, P.; Wang, W.; Watanabe, N.; Ziefle, G.; Kolditz, O.

    2016-12-01

    As part of the German site selection process for a high-level nuclear waste repository, different repository concepts in the geological candidate formations rock salt, clay stone and crystalline rock are being discussed. An open assessment of these concepts using numerical simulations requires physical models capturing the individual particularities of each rock type and associated geotechnical barrier concept to a comparable level of sophistication. In a joint work group of the Helmholtz Centre for Environmental Research (UFZ) and the German Federal Institute for Geosciences and Natural Resources (BGR), scientists of the UFZ are developing and implementing multiphysical process models while BGR scientists apply them to large scale analyses. The advances in simulation methods for waste repositories are incorporated into the open-source code OpenGeoSys. Here, recent application-driven progress in this context is highlighted. A robust implementation of visco-plasticity with temperature-dependent properties into a framework for the thermo-mechanical analysis of rock salt will be shown. The model enables the simulation of heat transport along with its consequences on the elastic response as well as on primary and secondary creep or the occurrence of dilatancy in the repository near field. Transverse isotropy, non-isothermal hydraulic processes and their coupling to mechanical stresses are taken into account for the analysis of repositories in clay stone. These processes are also considered in the near field analyses of engineered barrier systems, including the swelling/shrinkage of the bentonite material. The temperature-dependent saturation evolution around the heat-emitting waste container is described by different multiphase flow formulations. For all mentioned applications, we illustrate the workflow from model development and implementation, over verification and validation, to repository-scale application simulations using methods of high performance computing.

  14. The siting program of geological repository for spent fuel/high-level waste in Czech Republic

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Novotny, P.

    1993-12-31

    The management of high-level waste in Czech Republic have a very short history, because before the year 1989 spent nuclear fuel was re-exported back to USSR. The project ``Geological research of HLW repository in Czech Republic`` was initiated during 1990 by the Ministry of the Environment of the Czech Republic and by this project delegated the Czech Geological Survey (CGU) Prague. The first CGU project late in 1990 for multibarrier concept has proposed a geological repository to be located at a depth of about 500 m. Screening and studies of potential sites for repository started in 1991. First stage representedmore » regional siting of the Czech Republic for perspective rock types and massifs. In cooperation with GEOPHYSICS Co., Geophysical Institute of the Czech Academy of Sciences and Charles University Prague 27 perspective regions were selected, using criteria IAEA. This work in the Czech Republic was possible thanks to the detailed geological studies done in the past and thanks to the numerous archive data, concentrated in the central geological archive GEOFOND. Selection of perspective sites also respected natural conservation regions, regions conserving water and mineral waters resources. CGU opened up contact with countries with similar geological situation and started cooperation with SKB (Swedish Nuclear Fuel and Waste Management Co.). The Project of geological research for the next 10 years is a result of these activities.« less

  15. SEEK: a systems biology data and model management platform.

    PubMed

    Wolstencroft, Katherine; Owen, Stuart; Krebs, Olga; Nguyen, Quyen; Stanford, Natalie J; Golebiewski, Martin; Weidemann, Andreas; Bittkowski, Meik; An, Lihua; Shockley, David; Snoep, Jacky L; Mueller, Wolfgang; Goble, Carole

    2015-07-11

    Systems biology research typically involves the integration and analysis of heterogeneous data types in order to model and predict biological processes. Researchers therefore require tools and resources to facilitate the sharing and integration of data, and for linking of data to systems biology models. There are a large number of public repositories for storing biological data of a particular type, for example transcriptomics or proteomics, and there are several model repositories. However, this silo-type storage of data and models is not conducive to systems biology investigations. Interdependencies between multiple omics datasets and between datasets and models are essential. Researchers require an environment that will allow the management and sharing of heterogeneous data and models in the context of the experiments which created them. The SEEK is a suite of tools to support the management, sharing and exploration of data and models in systems biology. The SEEK platform provides an access-controlled, web-based environment for scientists to share and exchange data and models for day-to-day collaboration and for public dissemination. A plug-in architecture allows the linking of experiments, their protocols, data, models and results in a configurable system that is available 'off the shelf'. Tools to run model simulations, plot experimental data and assist with data annotation and standardisation combine to produce a collection of resources that support analysis as well as sharing. Underlying semantic web resources additionally extract and serve SEEK metadata in RDF (Resource Description Format). SEEK RDF enables rich semantic queries, both within SEEK and between related resources in the web of Linked Open Data. The SEEK platform has been adopted by many systems biology consortia across Europe. It is a data management environment that has a low barrier of uptake and provides rich resources for collaboration. This paper provides an update on the functions and features of the SEEK software, and describes the use of the SEEK in the SysMO consortium (Systems biology for Micro-organisms), and the VLN (virtual Liver Network), two large systems biology initiatives with different research aims and different scientific communities.

  16. Adaptable data management for systems biology investigations.

    PubMed

    Boyle, John; Rovira, Hector; Cavnor, Chris; Burdick, David; Killcoyne, Sarah; Shmulevich, Ilya

    2009-03-06

    Within research each experiment is different, the focus changes and the data is generated from a continually evolving barrage of technologies. There is a continual introduction of new techniques whose usage ranges from in-house protocols through to high-throughput instrumentation. To support these requirements data management systems are needed that can be rapidly built and readily adapted for new usage. The adaptable data management system discussed is designed to support the seamless mining and analysis of biological experiment data that is commonly used in systems biology (e.g. ChIP-chip, gene expression, proteomics, imaging, flow cytometry). We use different content graphs to represent different views upon the data. These views are designed for different roles: equipment specific views are used to gather instrumentation information; data processing oriented views are provided to enable the rapid development of analysis applications; and research project specific views are used to organize information for individual research experiments. This management system allows for both the rapid introduction of new types of information and the evolution of the knowledge it represents. Data management is an important aspect of any research enterprise. It is the foundation on which most applications are built, and must be easily extended to serve new functionality for new scientific areas. We have found that adopting a three-tier architecture for data management, built around distributed standardized content repositories, allows us to rapidly develop new applications to support a diverse user community.

  17. Push and pull models to manage patient consent and licensing of multimedia resources in digital repositories for case-based reasoning.

    PubMed

    Kononowicz, Andrzej A; Zary, Nabil; Davies, David; Heid, Jörn; Woodham, Luke; Hege, Inga

    2011-01-01

    Patient consents for distribution of multimedia constitute a significant element of medical case-based repositories in medicine. A technical challenge is posed by the right of patients to withdraw permission to disseminate their images or videos. A technical mechanism for spreading information about changes in multimedia usage licenses is sought. The authors gained their experience by developing and managing a large (>340 cases) repository of virtual patients within the European project eViP. The solution for dissemination of license status should reuse and extend existing metadata standards in medical education. Two methods: PUSH and PULL are described differing in the moment of update and the division of responsibilities between parties in the learning object exchange process. The authors recommend usage of the PUSH scenario because it is better adapted to legal requirements in many countries. It needs to be stressed that the solution is based on mutual trust of the exchange partners and therefore is most appropriate for use in educational alliances and consortia. It is hoped that the proposed models for exchanging consents and licensing information will become a crucial part of the technical frameworks for building case-based repositories.

  18. SNOMED CT module-driven clinical archetype management.

    PubMed

    Allones, J L; Taboada, M; Martinez, D; Lozano, R; Sobrido, M J

    2013-06-01

    To explore semantic search to improve management and user navigation in clinical archetype repositories. In order to support semantic searches across archetypes, an automated method based on SNOMED CT modularization is implemented to transform clinical archetypes into SNOMED CT extracts. Concurrently, query terms are converted into SNOMED CT concepts using the search engine Lucene. Retrieval is then carried out by matching query concepts with the corresponding SNOMED CT segments. A test collection of the 16 clinical archetypes, including over 250 terms, and a subset of 55 clinical terms from two medical dictionaries, MediLexicon and MedlinePlus, were used to test our method. The keyword-based service supported by the OpenEHR repository offered us a benchmark to evaluate the enhancement of performance. In total, our approach reached 97.4% precision and 69.1% recall, providing a substantial improvement of recall (more than 70%) compared to the benchmark. Exploiting medical domain knowledge from ontologies such as SNOMED CT may overcome some limitations of the keyword-based systems and thus improve the search experience of repository users. An automated approach based on ontology segmentation is an efficient and feasible way for supporting modeling, management and user navigation in clinical archetype repositories. Copyright © 2013 Elsevier Inc. All rights reserved.

  19. Analogues to features and processes of a high-level radioactive waste repository proposed for Yucca Mountain, Nevada

    USGS Publications Warehouse

    Simmons, Ardyth M.; Stuckless, John S.; with a Foreword by Abraham Van Luik, U.S. Department of Energy

    2010-01-01

    Natural analogues are defined for this report as naturally occurring or anthropogenic systems in which processes similar to those expected to occur in a nuclear waste repository are thought to have taken place over time periods of decades to millennia and on spatial scales as much as tens of kilometers. Analogues provide an important temporal and spatial dimension that cannot be tested by laboratory or field-scale experiments. Analogues provide one of the multiple lines of evidence intended to increase confidence in the safe geologic disposal of high-level radioactive waste. Although the work in this report was completed specifically for Yucca Mountain, Nevada, as the proposed geologic repository for high-level radioactive waste under the U.S. Nuclear Waste Policy Act, the applicability of the science, analyses, and interpretations is not limited to a specific site. Natural and anthropogenic analogues have provided and can continue to provide value in understanding features and processes of importance across a wide variety of topics in addressing the challenges of geologic isolation of radioactive waste and also as a contribution to scientific investigations unrelated to waste disposal. Isolation of radioactive waste at a mined geologic repository would be through a combination of natural features and engineered barriers. In this report we examine analogues to many of the various components of the Yucca Mountain system, including the preservation of materials in unsaturated environments, flow of water through unsaturated volcanic tuff, seepage into repository drifts, repository drift stability, stability and alteration of waste forms and components of the engineered barrier system, and transport of radionuclides through unsaturated and saturated rock zones.

  20. Tourism impacts of Three Mile Island and other adverse events: Implications for Lincoln County and other rural counties bisected by radioactive wastes intended for Yucca Mountain

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Himmelberger, J.J.; Ogneva-Himmelberger, Y.A.; Baughman, M.

    Whether the proposed Yucca Mountain nuclear waste repository system will adversely impact tourism in southern Nevada is an open question of particular importance to visitor-oriented rural counties bisected by planned waste transportation corridors (highway or rail). As part of one such county`s repository impact assessment program, tourism implications of Three Mile Island (TMI) and other major hazard events have been revisited to inform ongoing county-wide socioeconomic assessments and contingency planning efforts. This paper summarizes key research implications of such research as applied to Lincoln County, Nevada. Implications for other rural counties are discussed in light of the research findings. 29more » refs., 3 figs., 1 tab.« less

  1. Determination of In-situ Porosity and Investigation of Diffusion Processes at the Grimsel Test Site, Switzerland.

    NASA Astrophysics Data System (ADS)

    Biggin, C.; Ota, K.; Siittari-Kauppi, M.; Moeri, A.

    2004-12-01

    In the context of a repository for radioactive waste, 'matrix diffusion' is used to describe the process by which solute, flowing in distinct flow paths, penetrates the surrounding rock matrix. Diffusion into the matrix occurs in a connected system of pores or microfractures. Matrix diffusion provides a mechanism for greatly enlarging the area of rock surface in contact with advecting radionuclides, from that of the flow path surfaces (and infills), to a much larger portion of the bulk rock and increases the global pore volume which can retard radionuclides. In terms of a repository safety assessment, demonstration of a significant depth of diffusion-accessible pore space may result in a significant delay in the calculated release of any escaping radionuclides to the environment and a dramatic reduction in the resulting concentration released into the biosphere. For the last decade, Nagra has investigated in situ matrix diffusion at the Grimsel Test Site (GTS) in the Swiss Alps. The in situ investigations offer two distinct advantages to those performed in the lab, namely: 1. Lab-based determination of porosity and diffusivity can lead to an overestimation of matrix diffusion due to stress relief when the rock is sampled (which would overestimate the retardation in the geosphere) 2. Lab-based analysis usually examines small (cm scale) samples and cannot therefore account for any matrix heterogeneity over the hundreds or thousands of metres a typical flow path The in situ investigations described began with the Connected Porosity project, wherein a specially developed acrylic resin was injected into the rock matrix to fill the pore space and determine the depth of connected porosity. The resin was polymerised in situ and the entire rock mass removed by overcoring. The results indicated that lab-based porosity measurements may be two to three times higher than those obtained in situ. While the depth of accessible matrix from a water-conducting feature assumed in repository performance assessments is generally 1 to 10 cm, the results from the GTS in situ experiment suggested depths of several metres could be more appropriate. More recently, the Pore Space Geometry (PSG) experiment at the GTS has used a C-14 doped acrylic resin, combined with state-of-the-art digital beta autoradiography and fluorescence detection to examine a larger area of rock for determination of porosity and the degree of connected pore space. Analysis is currently ongoing and the key findings will be reported in this paper. Starting at the GTS in 2005, the Long-term Diffusion (LTD) project will investigate such processes over spatial and temporal scales more relevant to a repository than traditional lab-based experiments. In the framework of this experiment, long-term (10 to 50 years) in situ diffusion experiments and resin injection experiments are planned to verify current models for matrix diffusion as a radionuclide retardation process. This paper will discuss the findings of the first two experiments and their significance to repository safety assessments before discussing the strategy for the future in relation to the LTD project.

  2. Pretest characterization of WIPP experimental waste

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Johnson, J.; Davis, H.; Drez, P.E.

    The Waste Isolation Pilot Plant (WIPP) near Carlsbad, New Mexico, is an underground repository designed for the storage and disposal of transuranic (TRU) wastes from US Department of Energy (DOE) facilities across the country. The Performance Assessment (PA) studies for WIPP address compliance of the repository with applicable regulations, and include full-scale experiments to be performed at the WIPP site. These experiments are the bin-scale and alcove tests to be conducted by Sandia National Laboratories (SNL). Prior to conducting these experiments, the waste to be used in these tests needs to be characterized to provide data on the initial conditionsmore » for these experiments. This characterization is referred to as the Pretest Characterization of WIPP Experimental Waste, and is also expected to provide input to other programmatic efforts related to waste characterization. The purpose of this paper is to describe the pretest waste characterization activities currently in progress for the WIPP bin-scale waste, and to discuss the program plan and specific analytical protocols being developed for this characterization. The relationship between different programs and documents related to waste characterization efforts is also highlighted in this paper.« less

  3. Effects of microbial processes on gas generation under expected WIPP repository conditions: Annual report through 1992

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Francis, A.J.; Gillow, J.B.

    1993-09-01

    Microbial processes involved in gas generation from degradation of the organic constituents of transuranic waste under conditions expected at the Waste Isolation Pilot Plant (WIPP) repository are being investigated at Brookhaven National Laboratory. These laboratory studies are part of the Sandia National Laboratories -- WIPP Gas Generation Program. Gas generation due to microbial degradation of representative cellulosic waste was investigated in short-term (< 6 months) and long-term (> 6 months) experiments by incubating representative paper (filter paper, paper towels, and tissue) in WIPP brine under initially aerobic (air) and anaerobic (nitrogen) conditions. Samples from the WIPP surficial environment and undergroundmore » workings harbor gas-producing halophilic microorganisms, the activities of which were studied in short-term experiments. The microorganisms metabolized a variety of organic compounds including cellulose under aerobic, anaerobic, and denitrifying conditions. In long-term experiments, the effects of added nutrients (trace amounts of ammonium nitrate, phosphate, and yeast extract), no nutrients, and nutrients plus excess nitrate on gas production from cellulose degradation.« less

  4. Rolling Deck to Repository I: Designing a Database Infrastructure

    NASA Astrophysics Data System (ADS)

    Arko, R. A.; Miller, S. P.; Chandler, C. L.; Ferrini, V. L.; O'Hara, S. H.

    2008-12-01

    The NSF-supported academic research fleet collectively produces a large and diverse volume of scientific data, which are increasingly being shared across disciplines and contributed to regional and global syntheses. As both Internet connectivity and storage technology improve, it becomes practical for ships to routinely deliver data and documentation for a standard suite of underway instruments to a central shoreside repository. Routine delivery will facilitate data discovery and integration, quality assessment, cruise planning, compliance with funding agency and clearance requirements, and long-term data preservation. We are working collaboratively with ship operators and data managers to develop a prototype "data discovery system" for NSF-supported research vessels. Our goal is to establish infrastructure for a central shoreside repository, and to develop and test procedures for the routine delivery of standard data products and documentation to the repository. Related efforts are underway to identify tools and criteria for quality control of standard data products, and to develop standard interfaces and procedures for maintaining an underway event log. Development of a shoreside repository infrastructure will include: 1. Deployment and testing of a central catalog that holds cruise summaries and vessel profiles. A cruise summary will capture the essential details of a research expedition (operating institution, ports/dates, personnel, data inventory, etc.), as well as related documentation such as event logs and technical reports. A vessel profile will capture the essential details of a ship's installed instruments (manufacturer, model, serial number, reference location, etc.), with version control as the profile changes through time. The catalog's relational database schema will be based on the UNOLS Data Best Practices Committee's recommendations, and published as a formal XML specification. 2. Deployment and testing of a central repository that holds navigation and routine underway data. Based on discussion with ship operators and data managers at a workgroup meeting in September 2008, we anticipate that a subset of underway data could be delivered from ships to the central repository in near- realtime - enabling the integrated display of ship tracks at a public Web portal, for example - and a full data package could be delivered post-cruise by network transfer or disk shipment. Once ashore, data sets could be distributed to assembly centers such as the Shipboard Automated Meteorological and Oceanographic System (SAMOS) for routine processing, quality assessment, and synthesis efforts - as well as transmitted to national data centers such as NODC and NGDC for permanent archival. 3. Deployment and testing of a basic suite of Web services to make cruise summaries, vessel profiles, event logs, and navigation data easily available. A standard set of catalog records, maps, and navigation features will be published via the Open Archives Initiative (OAI) and Open Geospatial Consortium (OGC) protocols, which can then be harvested by partner data centers and/or embedded in client applications.

  5. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Perry, Frank Vinton; Kelley, Richard E.

    The DOE Spent Fuel and Waste Technology (SWFT) R&D Campaign is supporting research on crystalline rock, shale (argillite) and salt as potential host rocks for disposal of HLW and SNF in a mined geologic repository. The distribution of these three potential repository host rocks is limited to specific regions of the US and to different geologic and hydrologic environments (Perry et al., 2014), many of which may be technically suitable as a site for mined geologic disposal. This report documents a regional geologic evaluation of the Pierre Shale, as an example of evaluating a potentially suitable shale for siting amore » geologic HLW repository. This report follows a similar report competed in 2016 on a regional evaluation of crystalline rock that focused on the Superior Province of the north-central US (Perry et al., 2016).« less

  6. Science, institutional archives and open access: an overview and a pilot survey on the Italian cancer research institutions.

    PubMed

    Poltronieri, Elisabetta; Truccolo, Ivana; Di Benedetto, Corrado; Castelli, Mauro; Mazzocut, Mauro; Cognetti, Gaetana

    2010-12-20

    The Open Archive Initiative (OAI) refers to a movement started around the '90 s to guarantee free access to scientific information by removing the barriers to research results, especially those related to the ever increasing journal subscription prices. This new paradigm has reshaped the scholarly communication system and is closely connected to the build up of institutional repositories (IRs) conceived to the benefit of scientists and research bodies as a means to keep possession of their own literary production. The IRs are high-value tools which permit authors to gain visibility by enabling rapid access to scientific material (not only publications) thus increasing impact (citation rate) and permitting a multidimensional assessment of research findings. A survey was conducted in March 2010 to mainly explore the managing system in use for archiving the research finding adopted by the Italian Scientific Institutes for Research, Hospitalization and Health Care (IRCCS) of the oncology area within the Italian National Health Service (Servizio Sanitario Nazionale, SSN). They were asked to respond to a questionnaire intended to collect data about institutional archives, metadata formats and posting of full-text documents. The enquiry concerned also the perceived role of the institutional repository DSpace ISS, built up by the Istituto Superiore di Sanità (ISS) and based on a XML scheme for encoding metadata. Such a repository aims at acting as a unique reference point for the biomedical information produced by the Italian research institutions. An in-depth analysis has also been performed on the collection of information material addressed to patients produced by the institutions surveyed. The survey respondents were 6 out of 9. The results reveal the use of different practices and standard among the institutions concerning: the type of documentation collected, the software adopted, the use and format of metadata and the conditions of accessibility to the IRs. The Italian research institutions in the field of oncology are moving the first steps towards the philosophy of OA. The main effort should be the implementation of common procedures also in order to connect scientific publications to researchers curricula. In this framework, an important effort is represented by the project of ISS aimed to set a common interface able to allow migration of data from partner institutions to the OA compliant repository DSpace ISS.

  7. Science, institutional archives and open access: an overview and a pilot survey on the Italian cancer research institutions

    PubMed Central

    2010-01-01

    Background The Open Archive Initiative (OAI) refers to a movement started around the '90s to guarantee free access to scientific information by removing the barriers to research results, especially those related to the ever increasing journal subscription prices. This new paradigm has reshaped the scholarly communication system and is closely connected to the build up of institutional repositories (IRs) conceived to the benefit of scientists and research bodies as a means to keep possession of their own literary production. The IRs are high-value tools which permit authors to gain visibility by enabling rapid access to scientific material (not only publications) thus increasing impact (citation rate) and permitting a multidimensional assessment of research findings. Methods A survey was conducted in March 2010 to mainly explore the managing system in use for archiving the research finding adopted by the Italian Scientific Institutes for Research, Hospitalization and Health Care (IRCCS) of the oncology area within the Italian National Health Service (Servizio Sanitario Nazionale, SSN). They were asked to respond to a questionnaire intended to collect data about institutional archives, metadata formats and posting of full-text documents. The enquiry concerned also the perceived role of the institutional repository DSpace ISS, built up by the Istituto Superiore di Sanità (ISS) and based on a XML scheme for encoding metadata. Such a repository aims at acting as a unique reference point for the biomedical information produced by the Italian research institutions. An in-depth analysis has also been performed on the collection of information material addressed to patients produced by the institutions surveyed. Results The survey respondents were 6 out of 9. The results reveal the use of different practices and standard among the institutions concerning: the type of documentation collected, the software adopted, the use and format of metadata and the conditions of accessibility to the IRs. Conclusions The Italian research institutions in the field of oncology are moving the first steps towards the philosophy of OA. The main effort should be the implementation of common procedures also in order to connect scientific publications to researchers curricula. In this framework, an important effort is represented by the project of ISS aimed to set a common interface able to allow migration of data from partner institutions to the OA compliant repository DSpace ISS. PMID:21172002

  8. Enabling FAIR and Open Data - The Importance of Communities on Influencing Change

    NASA Astrophysics Data System (ADS)

    Stall, S.; Lehnert, K.; Robinson, E.; Parsons, M. A.; Hanson, B.; Cutcher-Gershenfeld, J.; Nosek, B.

    2017-12-01

    Our research ecosystem is diverse and dependent on many interacting stakeholders that influence and support the process of science. These include funders, institutions, libraries, publishers, researchers, data managers, repositories, archives and communities. Process improvement in this ecosystem thus usually needs support by more than one of these many stakeholders. For example, mandates for open data extend across this ecosystem. Solutions require these stakeholders to come together and agree upon improvements. Recently, the value of FAIR and Open Data has encouraged funders to sponsor discussions with tangible agreements that include the steps needed to move the ecosystem towards results. Work by many of these stakeholders over the past years have developed pilot efforts that are ready to be scaled with broader engagement. A partnership of the AGU, Earth Science Information Partners (ESIP), Research Data Alliance (RDA), Center for Open Science, and key publishers including Science, Nature, and the Proceedings of the National Academy of Science (PNAS) have agreed to work together to develop integrated processes, leveraging these pilots, to make FAIR and open data the default for Earth and space science publications. This effort will build on the work of COPDESS.org, ESIP, RDA, the scientific journals, and domain repositories to ensure that well documented data, preserved in a repository with community agreed-upon metadata, and supporting persistent identifiers becomes part of the expected research products submitted in support of each publication.

  9. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zheng, Liange; Rutqvist, Jonny; Kim, Kunhwi

    The focus of research within the UFD Campaign is on repository-induced interactions that may affect the key safety characteristics of an argillaceous rock. These include thermal-hydrological-mechanical-chemical (THMC) process interactions that occur as a result of repository construction and waste emplacement. Some of the key questions addressed in this report include the development of fracturing in the excavation damaged zone (EDZ) and THMC effects on the near-field argillaceous rock and buffer minerals and petrophysical characteristics, particularly the impacts of induced temperature rise caused by waste heat.

  10. The Telemetric and Holter ECG Warehouse Initiative (THEW): a Data Repository for the Design, Implementation and Validation of ECG-related Technologies

    PubMed Central

    Couderc, Jean-Philippe

    2011-01-01

    We present an initiative supported by the National Heart Lung, and Blood Institute and the Food and Drug Administration for the development of a repository containing continuous electrocardiographic information to be shared with the worldwide scientific community. We believe that sharing data reinforces open scientific inquiry. It encourages diversity of analysis and opinion while promoting new research and facilitating the education of new researchers. In this paper, we present the resources available in this initiative for the scientific community. We describe the set of ECG signals currently hosted and we briefly discuss the associated clinical information (medical history. Disease and study-specific endpoints) and software tools we propose. Currently, the repository contains more than 250GB of data from eight clinical studies including healthy individuals and cardiac patients. This data is available for the development, implementation and validation of technologies related to body-surface ECGs. To conclude, the Telemetric and Holter ECG Warehouse (THEW) is an initiative developed to benefit the scientific community and to advance the field of quantitative electrocardiography and cardiac safety. PMID:21097349

  11. Making proteomics data accessible and reusable: current state of proteomics databases and repositories.

    PubMed

    Perez-Riverol, Yasset; Alpi, Emanuele; Wang, Rui; Hermjakob, Henning; Vizcaíno, Juan Antonio

    2015-03-01

    Compared to other data-intensive disciplines such as genomics, public deposition and storage of MS-based proteomics, data are still less developed due to, among other reasons, the inherent complexity of the data and the variety of data types and experimental workflows. In order to address this need, several public repositories for MS proteomics experiments have been developed, each with different purposes in mind. The most established resources are the Global Proteome Machine Database (GPMDB), PeptideAtlas, and the PRIDE database. Additionally, there are other useful (in many cases recently developed) resources such as ProteomicsDB, Mass Spectrometry Interactive Virtual Environment (MassIVE), Chorus, MaxQB, PeptideAtlas SRM Experiment Library (PASSEL), Model Organism Protein Expression Database (MOPED), and the Human Proteinpedia. In addition, the ProteomeXchange consortium has been recently developed to enable better integration of public repositories and the coordinated sharing of proteomics information, maximizing its benefit to the scientific community. Here, we will review each of the major proteomics resources independently and some tools that enable the integration, mining and reuse of the data. We will also discuss some of the major challenges and current pitfalls in the integration and sharing of the data. © 2014 The Authors. PROTEOMICS published by Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  12. Data Collection, Collaboration, Analysis, and Publication Using the Open Data Repository's (ODR) Data Publisher

    NASA Astrophysics Data System (ADS)

    Lafuente, B.; Stone, N.; Bristow, T.; Keller, R. M.; Blake, D. F.; Downs, R. T.; Pires, A.; Dateo, C. E.; Fonda, M.

    2017-12-01

    In development for nearly four years, the Open Data Repository's (ODR) Data Publisher software has become a useful tool for researchers' data needs. Data Publisher facilitates the creation of customized databases with flexible permission sets that allow researchers to share data collaboratively while improving data discovery and maintaining ownership rights. The open source software provides an end-to-end solution from collection to final repository publication. A web-based interface allows researchers to enter data, view data, and conduct analysis using any programming language supported by JupyterHub (http://www.jupyterhub.org). This toolset makes it possible for a researcher to store and manipulate their data in the cloud from any internet capable device. Data can be embargoed in the system until a date selected by the researcher. For instance, open publication can be set to a date that coincides with publication of data analysis in a third party journal. In conjunction with teams at NASA Ames and the University of Arizona, a number of pilot studies are being conducted to guide the software development so that it allows them to publish and share their data. These pilots include (1) the Astrobiology Habitable Environments Database (AHED), a central searchable repository designed to promote and facilitate the integration and sharing of all the data generated by the diverse disciplines in astrobiology; (2) a database containing the raw and derived data products from the CheMin instrument on the MSL rover Curiosity (http://odr.io/CheMin), featuring a versatile graphing system, instructions and analytical tools to process the data, and a capability to download data in different formats; and (3) the Mineral Evolution project, which by correlating the diversity of mineral species with their ages, localities, and other measurable properties aims to understand how the episodes of planetary accretion and differentiation, plate tectonics, and origin of life lead to a selective evolution of mineral species through changes in temperature, pressure, and composition. Ongoing development will complete integration of third party meta-data standards and publishing data to the semantic web. This project is supported by the Science-Enabling Research Activity (SERA) and NASA NNX11AP82A, MSL.

  13. Process improvement methods increase the efficiency, accuracy, and utility of a neurocritical care research repository.

    PubMed

    O'Connor, Sydney; Ayres, Alison; Cortellini, Lynelle; Rosand, Jonathan; Rosenthal, Eric; Kimberly, W Taylor

    2012-08-01

    Reliable and efficient data repositories are essential for the advancement of research in Neurocritical care. Various factors, such as the large volume of patients treated within the neuro ICU, their differing length and complexity of hospital stay, and the substantial amount of desired information can complicate the process of data collection. We adapted the tools of process improvement to the data collection and database design of a research repository for a Neuroscience intensive care unit. By the Shewhart-Deming method, we implemented an iterative approach to improve the process of data collection for each element. After an initial design phase, we re-evaluated all data fields that were challenging or time-consuming to collect. We then applied root-cause analysis to optimize the accuracy and ease of collection, and to determine the most efficient manner of collecting the maximal amount of data. During a 6-month period, we iteratively analyzed the process of data collection for various data elements. For example, the pre-admission medications were found to contain numerous inaccuracies after comparison with a gold standard (sensitivity 71% and specificity 94%). Also, our first method of tracking patient admissions and discharges contained higher than expected errors (sensitivity 94% and specificity 93%). In addition to increasing accuracy, we focused on improving efficiency. Through repeated incremental improvements, we reduced the number of subject records that required daily monitoring from 40 to 6 per day, and decreased daily effort from 4.5 to 1.5 h/day. By applying process improvement methods to the design of a Neuroscience ICU data repository, we achieved a threefold improvement in efficiency and increased accuracy. Although individual barriers to data collection will vary from institution to institution, a focus on process improvement is critical to overcoming these barriers.

  14. Enhancing Ocean Research Data Access

    NASA Astrophysics Data System (ADS)

    Chandler, Cynthia; Groman, Robert; Shepherd, Adam; Allison, Molly; Arko, Robert; Chen, Yu; Fox, Peter; Glover, David; Hitzler, Pascal; Leadbetter, Adam; Narock, Thomas; West, Patrick; Wiebe, Peter

    2014-05-01

    The Biological and Chemical Oceanography Data Management Office (BCO-DMO) works in partnership with ocean science investigators to publish data from research projects funded by the Biological and Chemical Oceanography Sections and the Office of Polar Programs Antarctic Organisms & Ecosystems Program at the U.S. National Science Foundation. Since 2006, researchers have been contributing data to the BCO-DMO data system, and it has developed into a rich repository of data from ocean, coastal and Great Lakes research programs. While the ultimate goal of the BCO-DMO is to ensure preservation of NSF funded project data and to provide open access to those data, achievement of those goals is attained through a series of related phases that benefits from active collaboration and cooperation with a large community of research scientists as well as curators of data and information at complementary data repositories. The BCO-DMO is just one of many intermediate data management centers created to facilitate long-term preservation of data and improve access to ocean research data. Through partnerships with other data management professionals and active involvement in local and global initiatives, BCO-DMO staff members are working to enhance access to ocean research data available from the online BCO-DMO data system. Continuing efforts in use of controlled vocabulary terms, development of ontology design patterns and publication of content as Linked Open Data are contributing to improved discovery and availability of BCO-DMO curated data and increased interoperability of related content available from distributed repositories. We will demonstrate how Semantic Web technologies (e.g. RDF/XML, SKOS, OWL and SPARQL) have been integrated into BCO-DMO data access and delivery systems to better serve the ocean research community and to contribute to an expanding global knowledge network.

  15. An Optimal Centralized Carbon Dioxide Repository for Florida, USA

    PubMed Central

    Poiencot, Brandon; Brown, Christopher

    2011-01-01

    For over a decade, the United States Department of Energy, and engineers, geologists, and scientists from all over the world have investigated the potential for reducing atmospheric carbon emissions through carbon sequestration. Numerous reports exist analyzing the potential for sequestering carbon dioxide at various sites around the globe, but none have identified the potential for a statewide system in Florida, USA. In 2005, 83% of Florida’s electrical energy was produced by natural gas, coal, or oil (e.g., fossil fuels), from power plants spread across the state. In addition, only limited research has been completed on evaluating optimal pipeline transportation networks to centralized carbon dioxide repositories. This paper describes the feasibility and preliminary locations for an optimal centralized Florida-wide carbon sequestration repository. Linear programming optimization modeling is used to plan and route an idealized pipeline network to existing Florida power plants. Further analysis of the subsurface geology in these general locations will provide insight into the suitability of the subsurface conditions and the available capacity for carbon sequestration at selected possible repository sites. The identification of the most favorable site(s) is also presented. PMID:21695024

  16. An optimal centralized carbon dioxide repository for Florida, USA.

    PubMed

    Poiencot, Brandon; Brown, Christopher

    2011-04-01

    For over a decade, the United States Department of Energy, and engineers, geologists, and scientists from all over the world have investigated the potential for reducing atmospheric carbon emissions through carbon sequestration. Numerous reports exist analyzing the potential for sequestering carbon dioxide at various sites around the globe, but none have identified the potential for a statewide system in Florida, USA. In 2005, 83% of Florida's electrical energy was produced by natural gas, coal, or oil (e.g., fossil fuels), from power plants spread across the state. In addition, only limited research has been completed on evaluating optimal pipeline transportation networks to centralized carbon dioxide repositories. This paper describes the feasibility and preliminary locations for an optimal centralized Florida-wide carbon sequestration repository. Linear programming optimization modeling is used to plan and route an idealized pipeline network to existing Florida power plants. Further analysis of the subsurface geology in these general locations will provide insight into the suitability of the subsurface conditions and the available capacity for carbon sequestration at selected possible repository sites. The identification of the most favorable site(s) is also presented.

  17. mHealthApps: A Repository and Database of Mobile Health Apps.

    PubMed

    Xu, Wenlong; Liu, Yin

    2015-03-18

    The market of mobile health (mHealth) apps has rapidly evolved in the past decade. With more than 100,000 mHealth apps currently available, there is no centralized resource that collects information on these health-related apps for researchers in this field to effectively evaluate the strength and weakness of these apps. The objective of this study was to create a centralized mHealth app repository. We expect the analysis of information in this repository to provide insights for future mHealth research developments. We focused on apps from the two most established app stores, the Apple App Store and the Google Play Store. We extracted detailed information of each health-related app from these two app stores via our python crawling program, and then stored the information in both a user-friendly array format and a standard JavaScript Object Notation (JSON) format. We have developed a centralized resource that provides detailed information of more than 60,000 health-related apps from the Apple App Store and the Google Play Store. Using this information resource, we analyzed thousands of apps systematically and provide an overview of the trends for mHealth apps. This unique database allows the meta-analysis of health-related apps and provides guidance for research designs of future apps in the mHealth field.

  18. Genome resource banking for wildlife research, management, and conservation.

    PubMed

    Wildt, D E

    2000-01-01

    Cryobiology offers an important opportunity to assist in the management and study of wildlife, including endangered species. The benefits of developing genome resource banks for wildlife are profound, perhaps more so than for traditional uses in terms of livestock and human fertility. In addition to preserving heterozygosity and assisting in the genetic management of rare populations held in captivity, frozen repositories help insure wild populations against natural and human-induced catastrophes. Such banks also are an invaluable source of new knowledge (for basic and applied research) from thousands of species that have yet to be studied. However, it is crucial that genome resource banks for wildlife species be developed in a coordinated fashion that first benefits the conservation of biodiversity. Spurious collections will be of no advantage to genuine conservation. The Conservation Breeding Specialist Group (CBSG; of the International Union for the Conservation of Nature and Natural Resources' Species Survival Commission) has promoted international dialogue on this topic. CBSG working groups have recognized that such repositories be developed according to specific, scientific guidelines consistent with an international standard that ensures practicality, high-quality ethics, and cost-effectiveness. Areas requiring priority attention also are reviewed, including the need for more basic research, advocacy, and support for developing organized repositories of biomaterials representing the world's diverse biota.

  19. Research data management support for large-scale, long-term, interdisciplinary collaborative research centers with a focus on environmental sciences

    NASA Astrophysics Data System (ADS)

    Curdt, C.; Hoffmeister, D.; Bareth, G.; Lang, U.

    2017-12-01

    Science conducted in collaborative, cross-institutional research projects, requires active sharing of research ideas, data, documents and further information in a well-managed, controlled and structured manner. Thus, it is important to establish corresponding infrastructures and services for the scientists. Regular project meetings and joint field campaigns support the exchange of research ideas. Technical infrastructures facilitate storage, documentation, exchange and re-use of data as results of scientific output. Additionally, also publications, conference contributions, reports, pictures etc. should be managed. Both, knowledge and data sharing is essential to create synergies. Within the coordinated programme `Collaborative Research Center' (CRC), the German Research Foundation offers funding to establish research data management (RDM) infrastructures and services. CRCs are large-scale, interdisciplinary, multi-institutional, long-term (up to 12 years), university-based research institutions (up to 25 sub-projects). These CRCs address complex and scientifically challenging research questions. This poster presents the RDM services and infrastructures that have been established for two CRCs, both focusing on environmental sciences. Since 2007, a RDM support infrastructure and associated services have been set up for the CRC/Transregio 32 (CRC/TR32) `Patterns in Soil-Vegetation-Atmosphere-Systems: Monitoring, Modelling and Data Assimilation' (www.tr32.de). The experiences gained have been used to arrange RDM services for the CRC1211 `Earth - Evolution at the Dry Limit' (www.crc1211.de), funded since 2016. In both projects scientists from various disciplines collect heterogeneous data at field campaigns or by modelling approaches. To manage the scientific output, the TR32DB data repository (www.tr32db.de) has been designed and implemented for the CRC/TR32. This system was transferred and adapted to the CRC1211 needs (www.crc1211db.uni-koeln.de) in 2016. Both repositories support secure and sustainable data storage, backup, documentation, publication with DOIs, search, download, statistics as well as web mapping features. Moreover, RDM consulting and support services as well as training sessions are carried out regularly.

  20. NASA GSFC Tin Whisker Homepage http://nepp.nasa.gov/whisker

    NASA Technical Reports Server (NTRS)

    Shaw, Harry

    2000-01-01

    The NASA GSFC Tin Whisker Homepage provides general information and GSFC Code 562 experimentation results regarding the well known phenomenon of tin whisker formation from pure tin plated substrates. The objective of this www site is to provide a central repository for information pertaining to this phenomenon and to provide status of the GSFC experiments to understand the behavior of tin whiskers in space environments. The Tin Whisker www site is produced by Code 562. This www site does not provide information pertaining to patented or proprietary information. All of the information contained in this www site is at the level of that produced by industry and university researchers and is published at international conferences.

  1. Rolling Deck to Repository (R2R): Products and Services for the U.S. Research Fleet Community

    NASA Astrophysics Data System (ADS)

    Arko, R. A.; Carbotte, S. M.; Chandler, C. L.; Smith, S. R.; Stocks, K. I.

    2016-02-01

    The Rolling Deck to Repository (R2R) program is working to ensure open access to environmental sensor data routinely acquired by the U.S. academic research fleet. Currently 25 vessels deliver 7 TB/year of data to R2R from a suite of geophysical, oceanographic, meteorological, and navigational sensors on over 400 cruises worldwide. R2R ensures these data are preserved in trusted repositories, discoverable via standard protocols, and adequately documented for reuse. R2R has recently expanded to include the vessels Sikuliaq, operated by the University of Alaska; Falkor, operated by the Schmidt Ocean Institute; and Ronald H. Brown and Okeanos Explorer, operated by NOAA. R2R maintains a master catalog of U.S. research cruises, currently holding over 4,670 expeditions including vessel and cruise identifiers, start/end dates and ports, project titles and funding awards, science parties, dataset inventories with instrument types and file formats, data quality assessments, and links to related content at other repositories. Standard post-field cruise products are published including shiptrack navigation, near-real-time MET/TSG data, underway geophysical profiles, and CTD profiles. Software tools available to users include the R2R Event Logger and the R2R Nav Manager. A Digital Object Identifier (DOI) is published for each cruise, original field sensor dataset, standard post-field product, and document (e.g. cruise report) submitted by the science party. Scientists are linked to personal identifiers such as ORCIDs where available. Using standard identifiers such as DOIs and ORCIDs facilitates linking with journal publications and generation of citation metrics. R2R collaborates in the Ocean Data Interoperability Platform (ODIP) to strengthen links among regional and national data systems, populates U.S. cruises in the POGO global catalog, and is working toward membership in the DataONE alliance. It is a lead partner in the EarthCube GeoLink project, developing Semantic Web technologies to share data and documentation between repositories, and in the newly-launched EarthCube SeaView project, delivering data from R2R and other ocean data facilities to scientists using the Ocean Data View (ODV) software tool.

  2. Research Data Management Self-Education for Librarians: A Webliography

    ERIC Educational Resources Information Center

    Goben, Abigail; Raszewski, Rebecca

    2015-01-01

    As data as a scholarly object continues to grow in importance in the research community, librarians are undertaking increasing responsibilities regarding data management and curation. New library initiatives include assisting researchers in finding data sets for reuse; locating and hosting repositories for required archiving; consultations on…

  3. Ciênsação: Gaining a Feeling for Sciences

    ERIC Educational Resources Information Center

    de Oliveira, Marcos Henrique Abreu; Fischer, Robert

    2017-01-01

    Ciênsação, an open online repository for hands-on experiments, has been developed to convince teachers in Latin America that science is best experienced first hand. Permitting students to experiment autonomously in small groups can be a challenging endeavour for educators in these countries. We analyse the reasons that cause hesitation of teachers…

  4. The NCAR Digital Asset Services Hub (DASH): Implementing Unified Data Discovery and Access

    NASA Astrophysics Data System (ADS)

    Stott, D.; Worley, S. J.; Hou, C. Y.; Nienhouse, E.

    2017-12-01

    The National Center for Atmospheric Research (NCAR) Directorate created the Data Stewardship Engineering Team (DSET) to plan and implement an integrated single entry point for uniform digital asset discovery and access across the organization in order to improve the efficiency of access, reduce the costs, and establish the foundation for interoperability with other federated systems. This effort supports new policies included in federal funding mandates, NSF data management requirements, and journal citation recommendations. An inventory during the early planning stage identified diverse asset types across the organization that included publications, datasets, metadata, models, images, and software tools and code. The NCAR Digital Asset Services Hub (DASH) is being developed and phased in this year to improve the quality of users' experiences in finding and using these assets. DASH serves to provide engagement, training, search, and support through the following four nodes (see figure). DASH MetadataDASH provides resources for creating and cataloging metadata to the NCAR Dialect, a subset of ISO 19115. NMDEdit, an editor based on a European open source application, has been configured for manual entry of NCAR metadata. CKAN, an open source data portal platform, harvests these XML records (along with records output directly from databases) from a Web Accessible Folder (WAF) on GitHub for validation. DASH SearchThe NCAR Dialect metadata drives cross-organization search and discovery through CKAN, which provides the display interface of search results. DASH search will establish interoperability by facilitating metadata sharing with other federated systems. DASH ConsultingThe DASH Data Curation & Stewardship Coordinator assists with Data Management (DM) Plan preparation and advises on Digital Object Identifiers. The coordinator arranges training sessions on the DASH metadata tools and DM planning, and provides one-on-one assistance as requested. DASH RepositoryA repository is under development for NCAR datasets currently not in existing lab-managed archives. The DASH repository will be under NCAR governance and meet Trustworthy Repositories Audit & Certification (TRAC) requirements. This poster will highlight the processes, lessons learned, and current status of the DASH effort at NCAR.

  5. 76 FR 5390 - National Human Genome Research Institute; Notice of Closed Meeting

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-01-31

    ... Research Institute; Notice of Closed Meeting Pursuant to section 10(d) of the Federal Advisory Committee... unwarranted invasion of personal privacy. Place: National Human Genome Research Institute Special Emphasis Panel; NHGRI Sample Repository for Human Genetic Research. Date: March 3, 2011. Time: 1 p.m. to 2 p.m...

  6. Challenges and Best Practices for the Curation and Publication of Long-Tail Data with GFZ Data Services

    NASA Astrophysics Data System (ADS)

    Elger, Kirsten; Ulbricht, Damian; Bertelmann, Roland

    2017-04-01

    Open access to research data is an increasing international request and includes not only data underlying scholarly publication, but also raw and curated data. Especially in the framework of the observed shift in many scientific fields towards data science and data mining, data repositories are becoming important player as data archives and access point to curated research data. While general and institutional data repositories are available across all scientific disciplines, domain-specific data repositories are specialised for scientific disciplines, like, e.g., bio- or geosciences, with the possibility to use more discipline-specific and richer metadata models than general repositories. Data publication is increasingly regarded as important scientific achievement, and datasets with digital object identifier (DOI) are now fully citable in journal articles. Moreover, following in their signature of the "Statement of Commitment of the Coalition on Publishing Data in the Earth and Space Sciences" (COPDESS), many publishers have adopted their data policies and recommend and even request to store and publish data underlying scholarly publications in (domain-specific) data repositories and not as classical supplementary material directly attached to the respective article. The curation of large dynamic data from global networks in, e.g., seismology, magnetics or geodesy, always required a high grade of professional, IT-supported data management, simply to be able to store and access the huge number of files and manage dynamic datasets. In contrast to these, the vast amount of research data acquired by individual investigators or small teams known as 'long-tail data' was often not the focus for the development of data curation infrastructures. Nevertheless, even though they are small in size and highly variable, in total they represent a significant portion of the total scientific outcome. The curation of long-tail data requires more individual approaches and personal involvement of the data curator, especially regarding the data description. Here we will introduce best practices for the publication of long-tail data that are helping to reduce the individual effort, improve the quality of the data description. The data repository of GFZ Data Services, which is hosted at GFZ German Research Centre for Geosciences in Potsdam, is a domain-specific data repository for geosciences. In addition to large dynamic datasets from different disciplines, it has a large focus on the DOI-referenced publication of long-tail data with the aim to reach a high grade of reusability through a comprehensive data description and in the same time provide and distribute standardised, machine actionable metadata for data discovery (FAIR data). The development of templates for data reports, metadata provision by scientists via an XML Metadata Editor and discipline-specific DOI landing pages are helping both, the data curators to handle all kinds of datasets and enabling the scientists, i.e. user, to quickly decide whether a published dataset is fulfilling their needs. In addition, GFZ Data Services have developed DOI-registration services for several international networks (e.g. ICGEM, World Stress Map, IGETS, etc.). In addition, we have developed project-or network-specific designs of the DOI landing pages with the logo or design of the networks or project

  7. The tropical germplasm repository program at the USDA-ARS, Tropical Agriculture Research Station, Mayaguez, Puerto Rico

    USDA-ARS?s Scientific Manuscript database

    The USDA-ARS Tropical Agriculture Research Station is the only research entity within the National Plant Germplasm system in the insular Caribbean region. It houses germplasm collections of cultivated tropical/subtropical germplasm of bananas/plantains, cacao, mamey sapote, sapodilla, Spanish lime,...

  8. 48 CFR 1852.235-70 - Center for AeroSpace Information.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... avail itself of the services provided by the NASA Center for AeroSpace Information (CASI) (http://www.sti.nasa.gov) for the conduct of research or research and development required under this contract. CASI provides a variety of services and products as a NASA repository and database of research...

  9. 48 CFR 1852.235-70 - Center for AeroSpace Information.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... avail itself of the services provided by the NASA Center for AeroSpace Information (CASI) (http://www.sti.nasa.gov) for the conduct of research or research and development required under this contract. CASI provides a variety of services and products as a NASA repository and database of research...

  10. Minimum Information about a Spinal Cord Injury Experiment: A Proposed Reporting Standard for Spinal Cord Injury Experiments

    PubMed Central

    Ferguson, Adam R.; Popovich, Phillip G.; Xu, Xiao-Ming; Snow, Diane M.; Igarashi, Michihiro; Beattie, Christine E.; Bixby, John L.

    2014-01-01

    Abstract The lack of reproducibility in many areas of experimental science has a number of causes, including a lack of transparency and precision in the description of experimental approaches. This has far-reaching consequences, including wasted resources and slowing of progress. Additionally, the large number of laboratories around the world publishing articles on a given topic make it difficult, if not impossible, for individual researchers to read all of the relevant literature. Consequently, centralized databases are needed to facilitate the generation of new hypotheses for testing. One strategy to improve transparency in experimental description, and to allow the development of frameworks for computer-readable knowledge repositories, is the adoption of uniform reporting standards, such as common data elements (data elements used in multiple clinical studies) and minimum information standards. This article describes a minimum information standard for spinal cord injury (SCI) experiments, its major elements, and the approaches used to develop it. Transparent reporting standards for experiments using animal models of human SCI aim to reduce inherent bias and increase experimental value. PMID:24870067

  11. A national registry for juvenile dermatomyositis and other paediatric idiopathic inflammatory myopathies: 10 years' experience; the Juvenile Dermatomyositis National (UK and Ireland) Cohort Biomarker Study and Repository for Idiopathic Inflammatory Myopathies

    PubMed Central

    Martin, Neil; Krol, Petra; Smith, Sally; Murray, Kevin; Pilkington, Clarissa A.; Davidson, Joyce E.

    2011-01-01

    Objectives. The paediatric idiopathic inflammatory myopathies (IIMs) are a group of rare chronic inflammatory disorders of childhood, affecting muscle, skin and other organs. There is a severe lack of evidence base for current treatment protocols in juvenile myositis. The rarity of these conditions means that multicentre collaboration is vital to facilitate studies of pathogenesis, treatment and disease outcomes. We have established a national registry and repository for childhood IIM, which aims to improve knowledge, facilitate research and clinical trials, and ultimately to improve outcomes for these patients. Methods. A UK-wide network of centres and research group was established to contribute to the study. Standardized patient assessment, data collection forms and sample protocols were agreed. The Biobank includes collection of peripheral blood mononuclear cells, serum, genomic DNA and biopsy material. An independent steering committee was established to oversee the use of data/samples. Centre training was provided for patient assessment, data collection and entry. Results. Ten years after inception, the study has recruited 285 children, of which 258 have JDM or juvenile PM; 86% of the cases have contributed the biological samples. Serial sampling linked directly to the clinical database makes this a highly valuable resource. The study has been a platform for 20 sub-studies and attracted considerable funding support. Assessment of children with myositis in contributing centres has changed through participation in this study. Conclusions. This establishment of a multicentre registry and Biobank has facilitated research and contributed to progress in the management of a complex group of rare muscloskeletal conditions. PMID:20823094

  12. Role of natural analogs in performance assessment of nuclear waste repositories

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sagar, B.; Wittmeyer, G.W.

    1995-09-01

    Mathematical models of the flow of water and transport of radionuclides in porous media will be used to assess the ability of deep geologic repositories to safely contain nuclear waste. These models must, in some sense, be validated to ensure that they adequately describe the physical processes occurring within the repository and its geologic setting. Inasmuch as the spatial and temporal scales over which these models must be applied in performance assessment are very large, validation of these models against laboratory and small-scale field experiments may be considered inadequate. Natural analogs may provide validation data that are representative of physico-chemicalmore » processes that occur over spatial and temporal scales as large or larger than those relevant to repository design. The authors discuss the manner in which natural analog data may be used to increase confidence in performance assessment models and conclude that, while these data may be suitable for testing the basic laws governing flow and transport, there is insufficient control of boundary and initial conditions and forcing functions to permit quantitative validation of complex, spatially distributed flow and transport models. The authors also express their opinion that, for collecting adequate data from natural analogs, resources will have to be devoted to them that are much larger than are devoted to them at present.« less

  13. WFIRST: Data/Instrument Simulation Support at IPAC

    NASA Astrophysics Data System (ADS)

    Laine, Seppo; Akeson, Rachel; Armus, Lee; Bennett, Lee; Colbert, James; Helou, George; Kirkpatrick, J. Davy; Meshkat, Tiffany; Paladini, Roberta; Ramirez, Solange; Wang, Yun; Xie, Joan; Yan, Lin

    2018-01-01

    As part of WFIRST Science Center preparations, the IPAC Science Operations Center (ISOC) maintains a repository of 1) WFIRST data and instrument simulations, 2) tools to facilitate scientific performance and feasibility studies using the WFIRST, and 3) parameters summarizing the current design and predicted performance of the WFIRST telescope and instruments. The simulation repository provides access for the science community to simulation code, tools, and resulting analyses. Examples of simulation code with ISOC-built web-based interfaces include EXOSIMS (for estimating exoplanet yields in CGI surveys) and the Galaxy Survey Exposure Time Calculator. In the future the repository will provide an interface for users to run custom simulations of a wide range of coronagraph instrument (CGI) observations and sophisticated tools for designing microlensing experiments. We encourage those who are generating simulations or writing tools for exoplanet observations with WFIRST to contact the ISOC team so we can work with you to bring these to the attention of the broader astronomical community as we prepare for the exciting science that will be enabled by WFIRST.

  14. Parent perspectives on privacy and governance for a pediatric repository of non-biological, research data.

    PubMed

    Manhas, Kiran P; Page, Stacey; Dodd, Shawn X; Letourneau, Nicole; Ambrose, Aleta; Cui, Xinjie; Tough, Suzanne C

    2015-02-01

    Research data repositories (RDRs) are data storage entities where data can be submitted, stored, and subsequently accessed for purposes beyond the original intent. There is little information relating to non-biological RDRs, nor considerations regarding pediatric data storage and re-use. We examined parent perspectives on pediatric, non-biological RDRs. Qualitative, descriptive methods including both interviews and focus groups were used. Purposive sampling of adult participants in two provincial birth cohorts yielded 19 interviewees and 18 focus group participants (4 groups). Transcripts were analyzed by thematic content analysis. Parent research participants strongly supported the sharing of their own, and their child's, non-biological research data. Four themes emerged: that altruism has limits, that participants have ongoing privacy concerns, that some participants need the assurance of congruent values between themselves and researchers/research questions, and that opinions diverge for some governance issues. The establishment of RDRs is important and maximizes participants', researchers', and funders' investments. Participants as data donors have concerns relating to privacy, relationships, and governance that must be considered in RDR development. © The Author(s) 2014.

  15. Solutions for research data from a publisher's perspective

    NASA Astrophysics Data System (ADS)

    Cotroneo, P.

    2015-12-01

    Sharing research data has the potential to make research more efficient and reproducible. Elsevier has developed several initiatives to address the different needs of research data users. These include PANGEA Linked data, which provides geo-referenced, citable datasets from earth and life sciences, archived as supplementary data from publications by the PANGEA data repository; Mendeley Data, which allows users to freely upload and share their data; a database linking program that creates links between articles on ScienceDirect and datasets held in external data repositories such as EarthRef and EarthChem; a pilot for searching for research data through a map interface; an open data pilot that allows authors publishing in Elsevier journals to store and share research data and make this publicly available as a supplementary file alongside their article; and data journals, including Data in Brief, which allow researchers to share their data open access. Through these initiatives, researchers are not only encouraged to share their research data, but also supported in optimizing their research data management. By making data more readily citable and visible, and hence generating citations for authors, these initiatives also aim to ensure that researchers get the recognition they deserve for publishing their data.

  16. Extreme ground motions and Yucca Mountain

    USGS Publications Warehouse

    Hanks, Thomas C.; Abrahamson, Norman A.; Baker, Jack W.; Boore, David M.; Board, Mark; Brune, James N.; Cornell, C. Allin; Whitney, John W.

    2013-01-01

    Yucca Mountain is the designated site of the underground repository for the United States' high-level radioactive waste (HLW), consisting of commercial and military spent nuclear fuel, HLW derived from reprocessing of uranium and plutonium, surplus plutonium, and other nuclear-weapons materials. Yucca Mountain straddles the western boundary of the Nevada Test Site, where the United States has tested nuclear devices since the 1950s, and is situated in an arid, remote, and thinly populated region of Nevada, ~100 miles northwest of Las Vegas. Yucca Mountain was originally considered as a potential underground repository of HLW because of its thick units of unsaturated rocks, with the repository horizon being not only ~300 m above the water table but also ~300 m below the Yucca Mountain crest. The fundamental rationale for a geologic (underground) repository for HLW is to securely isolate these materials from the environment and its inhabitants to the greatest extent possible and for very long periods of time. Given the present climate conditions and what is known about the current hydrologic system and conditions around and in the mountain itself, one would anticipate that the rates of infiltration, corrosion, and transport would be very low—except for the possibility that repository integrity might be compromised by low-probability disruptive events, which include earthquakes, strong ground motion, and (or) a repository-piercing volcanic intrusion/eruption. Extreme ground motions (ExGM), as we use the phrase in this report, refer to the extremely large amplitudes of earthquake ground motion that arise at extremely low probabilities of exceedance (hazard). They first came to our attention when the 1998 probabilistic seismic hazard analysis for Yucca Mountain was extended to a hazard level of 10-8/yr (a 10-4/yr probability for a 104-year repository “lifetime”). The primary purpose of this report is to summarize the principal results of the ExGM research program as they have developed over the past 5 years; what follows will be focused on Yucca Mountain, but not restricted to it.

  17. Data publication, documentation and user friendly landing pages - improving data discovery and reuse

    NASA Astrophysics Data System (ADS)

    Elger, Kirsten; Ulbricht, Damian; Bertelmann, Roland

    2016-04-01

    Research data are the basis for scientific research and often irreplaceable (e.g. observational data). Storage of such data in appropriate, theme specific or institutional repositories is an essential part of ensuring their long term preservation and access. The free and open access to research data for reuse and scrutiny has been identified as a key issue by the scientific community as well as by research agencies and the public. To ensure the datasets to intelligible and usable for others they must be accompanied by comprehensive data description and standardized metadata for data discovery, and ideally should be published using digital object identifier (DOI). These make datasets citable and ensure their long-term accessibility and are accepted in reference lists of journal articles (http://www.copdess.org/statement-of-commitment/). The GFZ German Research Centre for Geosciences is the national laboratory for Geosciences in Germany and part of the Helmholtz Association, Germany's largest scientific organization. The development and maintenance of data systems is a key component of 'GFZ Data Services' to support state-of-the-art research. The datasets, archived in and published by the GFZ Data Repository cover all geoscientific disciplines and range from large dynamic datasets deriving from global monitoring seismic or geodetic networks with real-time data acquisition, to remotely sensed satellite products, to automatically generated data publications from a database for data from micro meteorological stations, to various model results, to geochemical and rock mechanical analyses from various labs, and field observations. The user-friendly presentation of published datasets via a DOI landing page is as important for reuse as the storage itself, and the required information is highly specific for each scientific discipline. If dataset descriptions are too general, or require the download of a dataset before knowing its suitability, many researchers often decide not to reuse a published dataset. In contrast to large data repositories without thematic specification, theme-specific data repositories have a large expertise in data discovery and opportunity to develop usable, discipline-specific formats and layouts for specific datasets, including consultation to different formats for the data description (e.g., via a Data Report or an article in a Data Journal) with full consideration of international metadata standards.

  18. UceWeb: a web-based collaborative tool for collecting and sharing quality of life data.

    PubMed

    Parimbelli, E; Sacchi, L; Rubrichi, S; Mazzanti, A; Quaglini, S

    2015-01-01

    This work aims at building a platform where quality-of-life data, namely utility coefficients, can be elicited not only for immediate use, but also systematically stored together with patient profiles to build a public repository to be further exploited in studies on specific target populations (e.g. cost/utility analyses). We capitalized on utility theory and previous experience to define a set of desirable features such a tool should show to facilitate sound elicitation of quality of life. A set of visualization tools and algorithms has been developed to this purpose. To make it easily accessible for potential users, the software has been designed as a web application. A pilot validation study has been performed on 20 atrial fibrillation patients. A collaborative platform, UceWeb, has been developed and tested. It implements the standard gamble, time trade-off and rating-scale utility elicitation methods. It allows doctors and patients to choose the mode of interaction to maximize patients’ comfort in answering difficult questions. Every utility elicitation may contribute to the growth of the repository. UceWeb can become a unique source of data allowing researchers both to perform more reliable comparisons among healthcare interventions and build statistical models to gain deeper insight into quality of life data.

  19. Operational Tsunami Modelling with TsunAWI for the German-Indonesian Tsunami Early Warning System: Recent Developments

    NASA Astrophysics Data System (ADS)

    Rakowsky, N.; Harig, S.; Androsov, A.; Fuchs, A.; Immerz, A.; Schröter, J.; Hiller, W.

    2012-04-01

    Starting in 2005, the GITEWS project (German-Indonesian Tsunami Early Warning System) established from scratch a fully operational tsunami warning system at BMKG in Jakarta. Numerical simulations of prototypic tsunami scenarios play a decisive role in a priori risk assessment for coastal regions and in the early warning process itself. Repositories with currently 3470 regional tsunami scenarios for GITEWS and 1780 Indian Ocean wide scenarios in support of Indonesia as a Regional Tsunami Service Provider (RTSP) were computed with the non-linear shallow water modell TsunAWI. It is based on a finite element discretisation, employs unstructured grids with high resolution along the coast and includes inundation. This contribution gives an overview on the model itself, the enhancement of the model physics, and the experiences gained during the process of establishing an operational code suited for thousands of model runs. Technical aspects like computation time, disk space needed for each scenario in the repository, or post processing techniques have a much larger impact than they had in the beginning when TsunAWI started as a research code. Of course, careful testing on artificial benchmarks and real events remains essential, but furthermore, quality control for the large number of scenarios becomes an important issue.

  20. What becomes of nuclear risk assessment in light of radiation hormesis?

    PubMed

    Cuttler, Jerry M

    2006-08-25

    A nuclear probabilistic risk or safety assessment (PRA or PSA) is a scientific calculation that uses assumptions and models to determine the likelihood of plant or fuel repository failures and the corresponding releases of radioactivity. Estimated radiation doses to the surrounding population are linked inappropriately to risks of cancer death and congenital malformations. Even though PRAs use very pessimistic assumptions, they demonstrate that nuclear power plants and fuel repositories are very safe compared with the health risks of other generating options or other risks that people readily accept. Because of the frightening negative images and the exaggerated safety and health concerns that are communicated, many people judge nuclear risks to be unacceptable and do not favour nuclear plants. Large-scale tests and experience with nuclear accidents demonstrate that even severe accidents expose the public to only low doses of radiation, and a century of research has demonstrated that such exposures are beneficial to health. A scientific basis for this phenomenon now exists. PRAs are valuable tools for improving plant designs, but if nuclear power is to play a significant role in meeting future energy needs, we must communicate its many real benefits and dispel the negative images formed by unscientific extrapolations of harmful effects at high doses.

  1. Cigeo, the French Geological Repository Project - 13022

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Labalette, Thibaud; Harman, Alain; Dupuis, Marie-Claude

    The Cigeo industrial-scale geological disposal centre is designed for the disposal of the most highly-radioactive French waste. It will be built in an argillite formation of the Callovo-Oxfordian dating back 160 million years. The Cigeo project is located near the Bure village in the Paris Basin. The argillite formation was studied since 1974, and from the Meuse/Haute-Marne underground research laboratory since end of 1999. Most of the waste to be disposed of in the Cigeo repository comes from nuclear power plants and from reprocessing of their spent fuel. (authors)

  2. Review of DOE Waste Package Program. Semiannual report, October 1984-March 1985. Volume 8

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Davis, M.S.

    1985-12-01

    A large number of technical reports on waste package component performance were reviewed over the last year in support of the NRC`s review of the Department of Energy`s (DOE`s) Environmental Assessment reports. The intent was to assess in some detail the quantity and quality of the DOE data and their relevance to the high-level waste repository site selection process. A representative selection of the reviews is presented for the salt, basalt, and tuff repository projects. Areas for future research have been outlined. 141 refs.

  3. A Digital Repository and Execution Platform for Interactive Scholarly Publications in Neuroscience.

    PubMed

    Hodge, Victoria; Jessop, Mark; Fletcher, Martyn; Weeks, Michael; Turner, Aaron; Jackson, Tom; Ingram, Colin; Smith, Leslie; Austin, Jim

    2016-01-01

    The CARMEN Virtual Laboratory (VL) is a cloud-based platform which allows neuroscientists to store, share, develop, execute, reproduce and publicise their work. This paper describes new functionality in the CARMEN VL: an interactive publications repository. This new facility allows users to link data and software to publications. This enables other users to examine data and software associated with the publication and execute the associated software within the VL using the same data as the authors used in the publication. The cloud-based architecture and SaaS (Software as a Service) framework allows vast data sets to be uploaded and analysed using software services. Thus, this new interactive publications facility allows others to build on research results through reuse. This aligns with recent developments by funding agencies, institutions, and publishers with a move to open access research. Open access provides reproducibility and verification of research resources and results. Publications and their associated data and software will be assured of long-term preservation and curation in the repository. Further, analysing research data and the evaluations described in publications frequently requires a number of execution stages many of which are iterative. The VL provides a scientific workflow environment to combine software services into a processing tree. These workflows can also be associated with publications and executed by users. The VL also provides a secure environment where users can decide the access rights for each resource to ensure copyright and privacy restrictions are met.

  4. Scoping review and evaluation of SMS/text messaging platforms for mHealth projects or clinical interventions.

    PubMed

    Iribarren, Sarah J; Brown, William; Giguere, Rebecca; Stone, Patricia; Schnall, Rebecca; Staggers, Nancy; Carballo-Diéguez, Alex

    2017-05-01

    Mobile technology supporting text messaging interventions (TMIs) continues to evolve, presenting challenges for researchers and healthcare professionals who need to choose software solutions to best meet their program needs. The objective of this review was to systematically identify and compare text messaging platforms and to summarize their advantages and disadvantages as described in peer-reviewed literature. A scoping review was conducted using four steps: 1) identify currently available platforms through online searches and in mHealth repositories; 2) expand evaluation criteria of an mHealth mobile messaging toolkit and integrate prior user experiences as researchers; 3) evaluate each platform's functions and features based on the expanded criteria and a vendor survey; and 4) assess the documentation of platform use in the peer-review literature. Platforms meeting inclusion criteria were assessed independently by three reviewers and discussed until consensus was reached. The PRISMA guidelines were followed to report findings. Of the 1041 potentially relevant search results, 27 platforms met inclusion criteria. Most were excluded because they were not platforms (e.g., guides, toolkits, reports, or SMS gateways). Of the 27 platforms, only 12 were identified in existing mHealth repositories, 10 from Google searches, while five were found in both. The expanded evaluation criteria included 22 items. Results indicate no uniform presentation of platform features and functions, often making these difficult to discern. Fourteen of the platforms were reported as open source, 10 focused on health care and 16 were tailored to meet needs of low resource settings (not mutually exclusive). Fifteen platforms had do-it-yourself setup (programming not required) while the remainder required coding/programming skills or setups could be built to specification by the vendor. Frequently described features included data security and access to the platform via cloud-based systems. Pay structures and reported targeted end-users varied. Peer-reviewed publications listed only 6 of the 27 platforms across 21 publications. The majority of these articles reported the name of the platform used but did not describe advantages or disadvantages. Searching for and comparing mHealth platforms for TMIs remains a challenge. The results of this review can serve as a resource for researchers and healthcare professionals wanting to integrate TMIs into health interventions. Steps to identify, compare and assess advantages and disadvantages are outlined for consideration. Expanded evaluation criteria can be used by future researchers. Continued and more comprehensive platform tools should be integrated into mHealth repositories. Detailed descriptions of platform advantages and disadvantages are needed when mHealth researchers publish findings to expand the body of research on TMI tools for healthcare. Standardized descriptions and features are recommended for vendor sites. Copyright © 2017 Elsevier B.V. All rights reserved.

  5. Scoping Review and Evaluation of SMS/text Messaging Platforms for mHealth Projects or Clinical Interventions

    PubMed Central

    Iribarren, Sarah; Brown, William; Giguere, Rebecca; Stone, Patricia; Schnall, Rebecca; Staggers, Nancy; Carballo-Diéguez, Alex

    2017-01-01

    Objectives Mobile technology supporting text messaging interventions (TMIs) continues to evolve, presenting challenges for researchers and healthcare professionals who need to choose software solutions to best meet their program needs. The objective of this review was to systematically identify and compare text messaging platforms and to summarize their advantages and disadvantages as described in peer-reviewed literature. Methods A scoping review was conducted using four steps: 1) identify currently available platforms through online searches and in mHealth repositories; 2) expand evaluation criteria of an mHealth mobile messaging toolkit and prior user experiences as researchers; 3) evaluate each platform’s functions and features based on the expanded criteria and a vendor survey; and 4) assess the documentation of platform use in the peer-review literature. Platforms meeting inclusion criteria were assessed independently by three reviewers and discussed until consensus was reached. The PRISMA guidelines were followed to report findings. Results Of the 1041 potentially relevant search results, 27 platforms met inclusion criteria. Most were excluded because they were not platforms (e.g., guides, toolkits, reports, or SMS gateways). Of the 27 platforms, only 12 were identified in existing mHealth repositories, 10 from Google searches, while five were found in both. The expanded evaluation criteria included 22 items. Results indicate no uniform presentation of platform features and functions, often making these difficult to discern. Fourteen of the platforms were reported as open source, 10 focused on health care and 16 were tailored to meet needs of low resource settings (not mutually exclusive). Fifteen platforms had do-it-yourself setup (programming not required) while the remainder required coding/programming skills or setups could be built to specification by the vendor. Frequently described features included data security and access to the platform via cloud-based systems. Pay structures and reported targeted end-users varied. Peer-reviewed publications listed only 6 of the 27 platforms across 21 publications. The majority of these articles reported the name of the platform used but did not describe advantages or disadvantages. Conclusions Searching for and comparing mHealth platforms for TMIs remains a challenge. The results of this review can serve as a resource for researchers and healthcare professionals wanting to integrate TMIs into health interventions. Steps to identify, compare and assess advantages and disadvantages are outlined for consideration. Expanded evaluation criteria can be used by future researchers. Continued and more comprehensive platform tools should be integrated into mHealth repositories. Detailed descriptions of platform advantages and disadvantages are needed when mHealth researchers publish findings to expand the body of research on texting-based tools for healthcare. Standardized descriptions and features are recommended for vendor sites. PMID:28347445

  6. Adaptable data management for systems biology investigations

    PubMed Central

    Boyle, John; Rovira, Hector; Cavnor, Chris; Burdick, David; Killcoyne, Sarah; Shmulevich, Ilya

    2009-01-01

    Background Within research each experiment is different, the focus changes and the data is generated from a continually evolving barrage of technologies. There is a continual introduction of new techniques whose usage ranges from in-house protocols through to high-throughput instrumentation. To support these requirements data management systems are needed that can be rapidly built and readily adapted for new usage. Results The adaptable data management system discussed is designed to support the seamless mining and analysis of biological experiment data that is commonly used in systems biology (e.g. ChIP-chip, gene expression, proteomics, imaging, flow cytometry). We use different content graphs to represent different views upon the data. These views are designed for different roles: equipment specific views are used to gather instrumentation information; data processing oriented views are provided to enable the rapid development of analysis applications; and research project specific views are used to organize information for individual research experiments. This management system allows for both the rapid introduction of new types of information and the evolution of the knowledge it represents. Conclusion Data management is an important aspect of any research enterprise. It is the foundation on which most applications are built, and must be easily extended to serve new functionality for new scientific areas. We have found that adopting a three-tier architecture for data management, built around distributed standardized content repositories, allows us to rapidly develop new applications to support a diverse user community. PMID:19265554

  7. 75 FR 36427 - National Heart, Lung, and Blood Institute; Notice of Closed Meetings

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-06-25

    ... Cardiovascular Disease. Date: July 29, 2010. Time: 10 a.m. to 3 p.m. Agenda: To review and evaluate grant... Special Emphasis Panel, Resource Related Research Project in Lung Disease BioRepository. Date: July 15... Disorders Research; 93.837, Heart and Vascular Diseases Research; 93.838, Lung Diseases Research; 93.839...

  8. KiMoSys: a web-based repository of experimental data for KInetic MOdels of biological SYStems

    PubMed Central

    2014-01-01

    Background The kinetic modeling of biological systems is mainly composed of three steps that proceed iteratively: model building, simulation and analysis. In the first step, it is usually required to set initial metabolite concentrations, and to assign kinetic rate laws, along with estimating parameter values using kinetic data through optimization when these are not known. Although the rapid development of high-throughput methods has generated much omics data, experimentalists present only a summary of obtained results for publication, the experimental data files are not usually submitted to any public repository, or simply not available at all. In order to automatize as much as possible the steps of building kinetic models, there is a growing requirement in the systems biology community for easily exchanging data in combination with models, which represents the main motivation of KiMoSys development. Description KiMoSys is a user-friendly platform that includes a public data repository of published experimental data, containing concentration data of metabolites and enzymes and flux data. It was designed to ensure data management, storage and sharing for a wider systems biology community. This community repository offers a web-based interface and upload facility to turn available data into publicly accessible, centralized and structured-format data files. Moreover, it compiles and integrates available kinetic models associated with the data. KiMoSys also integrates some tools to facilitate the kinetic model construction process of large-scale metabolic networks, especially when the systems biologists perform computational research. Conclusions KiMoSys is a web-based system that integrates a public data and associated model(s) repository with computational tools, providing the systems biology community with a novel application facilitating data storage and sharing, thus supporting construction of ODE-based kinetic models and collaborative research projects. The web application implemented using Ruby on Rails framework is freely available for web access at http://kimosys.org, along with its full documentation. PMID:25115331

  9. Rolling Deck to Repository (R2R): Technical Design - Experiences and Lessons (Invited)

    NASA Astrophysics Data System (ADS)

    Arko, R. A.; Carbotte, S. M.; Miller, S. P.; Chandler, C. L.; Ferrini, V.; Stocks, K.; Maffei, A. R.; Smith, S. R.; Bourassa, M. A.; McLean, S. J.; Alberts, J. C.

    2009-12-01

    The NSF-funded Rolling Deck to Repository (R2R) project envisions the academic research fleet as an integrated global observing system, with routine “underway” sensor data flowing directly from research vessels to a central shore-side repository. It is a complex endeavor involving many stakeholders - technicians at sea, data managers on shore, ship schedulers, clearance officers, funding agencies, National Data Centers, data synthesis projects, the science community, and the public - working toward a common goal of acquiring, documenting, archiving, evaluating, and disseminating high-quality scientific data. The technical design for R2R is guided by several key principles: 1) The data pipeline is modular, so that initial stages (e.g. inventory and review of data shipments, posting of catalog records and track maps) may proceed routinely for every cruise, while later stages (e.g. quality assessment and production of file-level metadata) may proceed at different rates for different data types; 2) Cruise documentation (e.g. sailing orders, review/release of data inventories, vessel profiles) is gathered primarily via an authenticated Web portal, linked with the UNOLS scheduling database to synchronize vocabularies and eliminate redundancies; and 3) Every data set will be documented and delivered to the appropriate National Data Center for long-term archiving and dissemination after proprietary holds are cleared, while R2R maintains a master cruise catalog that links all the data sets together. This design accommodates the diversity of instrument types, data volumes, and shipment schedules among fleet operators. During its pilot development period, R2R has solicited feedback at community workshops, UNOLS meetings, and conference presentations, including fleet-wide surveys of current practices and instrument inventories. Several vessel operators began submitting cruise data and documentation during the pilot, providing a test bed for database development and Web portal design as well as feedback on delivery formats and data policies. Visits to operating institutions, including time at sea, have been critical to understanding the full range of vessel classes, capabilities, and concerns, and will continue to be an integral component of the R2R project.

  10. Native American Art and Culture: Documentary Resources.

    ERIC Educational Resources Information Center

    Lawrence, Deirdre

    1992-01-01

    Presents a brief overview of the evolution of documentary material of Native American cultures and problems confronted by researchers in locating relevant information. Bibliographic sources for research are discussed and a directory of major repositories of Native American art documentation is provided. (EA)

  11. 10 CFR 961.11 - Text of the contract.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... characteristic, of a specific or technical nature. It may, for example, document research, experimental... computer software documentation). Examples of technical data include research and engineering data... repository, to take title to the spent nuclear fuel or high-level radioactive waste involved as expeditiously...

  12. 10 CFR 961.11 - Text of the contract.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... characteristic, of a specific or technical nature. It may, for example, document research, experimental... computer software documentation). Examples of technical data include research and engineering data... repository, to take title to the spent nuclear fuel or high-level radioactive waste involved as expeditiously...

  13. Leveraging of Open EMR Architecture for Clinical Trial Accrual

    PubMed Central

    Afrin, Lawrence B.; Oates, James C.; Boyd, Caroline K.; Daniels, Mark S.

    2003-01-01

    Accrual to clinical trials is a major bottleneck in scientific progress in clinical medicine. Many methods for identifying potential subjects and improving accrual have been pursued; few have succeeded, and none have proven generally reproducible or scalable. We leveraged the open architecture of the core clinical data repository of our electronic medical record system to prototype a solution for this problem in a manner consistent with contemporary regulations and research ethics. We piloted the solution with a local investigator-initiated trial for which candidate identification was expected to be difficult. Key results in the eleven months of experience to date include automated screening of 7,296,708 lab results from 69,288 patients, detection of 1,768 screening tests of interest, identification of 70 potential candidates who met all further automated criteria, and accrual of three candidates to the trial. Hypotheses for this disappointing impact on accrual, and directions for future research, are discussed. PMID:14728125

  14. Optimizing Data Center Services to Foster Stewardship and Use of Geospatial Data by Heterogeneous Populations of Users

    NASA Astrophysics Data System (ADS)

    Downs, R. R.; Chen, R. S.; de Sherbinin, A. M.

    2017-12-01

    Growing recognition of the importance of sharing scientific data more widely and openly has refocused attention on the state of data repositories, including both discipline- or topic-oriented data centers and institutional repositories. Data creators often have several alternatives for depositing and disseminating their natural, social, health, or engineering science data. In selecting a repository for their data, data creators and other stakeholders such as their funding agencies may wish to consider the user community or communities served, the type and quality of data products already offered, and the degree of data stewardship and associated services provided. Some data repositories serve general communities, e.g., those in their host institution or region, whereas others tailor their services to particular scientific disciplines or topical areas. Some repositories are selective when acquiring data and conduct extensive curation and reviews to ensure that data products meet quality standards. Many repositories have secured credentials and established a track record for providing trustworthy, high quality data and services. The NASA Socioeconomic Data and Applications Center (SEDAC) serves users interested in human-environment interactions, including researchers, students, and applied users from diverse sectors. SEDAC is selective when choosing data for dissemination, conducting several reviews of data products and services prior to release. SEDAC works with data producers to continually improve the quality of its open data products and services. As a Distributed Active Archive Center (DAAC) of the NASA Earth Observing System Data and Information System, SEDAC is committed to improving the accessibility, interoperability, and usability of its data in conjunction with data available from other DAACs, as well as other relevant data sources. SEDAC is certified as a Regular Member of the International Council for Science World Data System (ICSU-WDS).

  15. Evaluation of geotechnical monitoring data from the ESF North Ramp Starter Tunnel, April 1994 to June 1995. Revision 1

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    1995-11-01

    This report presents the results of instrumentation measurements and observations made during construction of the North Ramp Starter Tunnel (NRST) of the Exploratory Studies Facility (ESF). The information in this report was developed as part of the Design Verification Study, Section 8.3.1.15.1.8 of the Yucca Mountain Site Characterization Plan (DOE 1988). The ESF is being constructed by the US Department of Energy (DOE) to evaluate the feasibility of locating a potential high-level nuclear waste repository on lands within and adjacent to the Nevada Test Site (NTS), Nye County, Nevada. The Design Verification Studies are performed to collect information during constructionmore » of the ESF that will be useful for design and construction of the potential repository. Four experiments make up the Design Verification Study: Evaluation of Mining Methods, Monitoring Drift Stability, Monitoring of Ground Support Systems, and The Air Quality and Ventilation Experiment. This report describes Sandia National Laboratories` (SNL) efforts in the first three of these experiments in the NRST.« less

  16. The Function Biomedical Informatics Research Network Data Repository

    PubMed Central

    Keator, David B.; van Erp, Theo G.M.; Turner, Jessica A.; Glover, Gary H.; Mueller, Bryon A.; Liu, Thomas T.; Voyvodic, James T.; Rasmussen, Jerod; Calhoun, Vince D.; Lee, Hyo Jong; Toga, Arthur W.; McEwen, Sarah; Ford, Judith M.; Mathalon, Daniel H.; Diaz, Michele; O’Leary, Daniel S.; Bockholt, H. Jeremy; Gadde, Syam; Preda, Adrian; Wible, Cynthia G.; Stern, Hal S.; Belger, Aysenil; McCarthy, Gregory; Ozyurt, Burak; Potkin, Steven G.

    2015-01-01

    The Function Biomedical Informatics Research Network (FBIRN) developed methods and tools for conducting multi-scanner functional magnetic resonance imaging (fMRI) studies. Method and tool development were based on two major goals: 1) to assess the major sources of variation in fMRI studies conducted across scanners, including instrumentation, acquisition protocols, challenge tasks, and analysis methods, and 2) to provide a distributed network infrastructure and an associated federated database to host and query large, multi-site, fMRI and clinical datasets. In the process of achieving these goals the FBIRN test bed generated several multi-scanner brain imaging data sets to be shared with the wider scientific community via the BIRN Data Repository (BDR). The FBIRN Phase 1 dataset consists of a traveling subject study of 5 healthy subjects, each scanned on 10 different 1.5 to 4 Tesla scanners. The FBIRN Phase 2 and Phase 3 datasets consist of subjects with schizophrenia or schizoaffective disorder along with healthy comparison subjects scanned at multiple sites. In this paper, we provide concise descriptions of FBIRN’s multi-scanner brain imaging data sets and details about the BIRN Data Repository instance of the Human Imaging Database (HID) used to publicly share the data. PMID:26364863

  17. Digital data preservation for scholarly publications in astronomy

    NASA Astrophysics Data System (ADS)

    Choudhury, Sayeed; di Lauro, Tim; Szalay, Alex; Vishniac, Ethan; Hanisch, Robert; Steffen, Julie; Milkey, Robert; Ehling, Teresa; Plante, Ray

    2007-11-01

    Astronomy is similar to other scientific disciplines in that scholarly publication relies on the presentation and interpretation of data. But although astronomy now has archives for its primary research telescopes and associated surveys, the highly processed data that is presented in the peer-reviewed journals and is the basis for final analysis and interpretation is generally not archived and has no permanent repository. We have initiated a project whose goal is to implement an end-to-end prototype system which, through a partnership of a professional society, that society's scholarly publications/publishers, research libraries, and an information technology substrate provided by the Virtual Observatory, will capture high-level digital data as part of the publication process and establish a distributed network of curated, permanent data repositories. The data in this network will be accessible through the research journals, astronomy data centers, and Virtual Observatory data discovery portals.

  18. [The Open Access Initiative (OAI) in the scientific literature].

    PubMed

    Sánchez-Martín, Francisco M; Millán Rodríguez, Félix; Villavicencio Mavrich, Humberto

    2009-01-01

    According to the declaration of the Budapest Open Access Initiative (OAI) is defined as a editorial model in which access to scientific journal literature and his use are free. Free flow of information allowed by Internet has been the basis of this initiative. The Bethesda and the Berlin declarations, supported by some international agencies, proposes to require researchers to deposit copies of all articles published in a self-archive or an Open Access repository, and encourage researchers to publish their research papers in journals Open Access. This paper reviews the keys of the OAI, with their strengths and controversial aspects; and it discusses the position of databases, search engines and repositories of biomedical information, as well as the attitude of the scientists, publishers and journals. So far the journal Actas Urológicas Españolas (Act Urol Esp) offer their contents on Open Access as On Line in Spanish and English.

  19. GaitaBase: Web-based repository system for gait analysis.

    PubMed

    Tirosh, Oren; Baker, Richard; McGinley, Jenny

    2010-02-01

    The need to share gait analysis data to improve clinical decision support has been recognised since the early 1990s. GaitaBase has been established to provide a web-accessible repository system of gait analysis data to improve the sharing of data across local and international clinical and research community. It is used by several clinical and research groups across the world providing cross-group access permissions to retrieve and analyse the data. The system is useful for bench-marking and quality assurance, clinical consultation, and collaborative research. It has the capacity to increase the population sample size and improve the quality of 'normative' gait data. In addition the accumulated stored data may facilitate clinicians in comparing their own gait data with others, and give a valuable insight into how effective specific interventions have been for others. 2009 Elsevier Ltd. All rights reserved.

  20. Physico-chemical interactions at the concrete-bitumen interface of nuclear waste repositories

    NASA Astrophysics Data System (ADS)

    Bertron, A.; Ranaivomanana, H.; Jacquemet, N.; Erable, B.; Sablayrolles, C.; Escadeillas, G.; Albrecht, A.

    2013-07-01

    This study investigates the fate of nitrate and organic acids at the bitumenconcrete-steel interface within a repository storage cell for long-lived, intermediatelevel, radioactive wastes. The interface was simulated by a multiphase system in which cementitious matrices (CEM V-paste specimens) were exposed to bitumen model leachates consisting of nitrates and acetic acid with and without oxalic acid, chemical compounds likely to be released by bitumen. Leaching experiments were conducted with daily renewal of the solutions in order to accelerate reactions. C-steel chips, simulating the presence of steel in the repository, were added in the systems for some experiments. The concentrations of anions (acetate, oxalate, nitrate, and nitrite) and cations (calcium, potassium, ammonium) and the pH were monitored over time. Mineralogical changes of the cementitious matrices were analysed by XRD. The results confirmed the stability of nitrates in the absence of steel, whereas, reduction of nitrates was observed in the presence of steel (production of NH4+). The action of acetic acid on the cementitious matrix was similar to that of ordinary leaching; no specific interaction was detected between acetate and cementitious cations. The reaction of oxalic acid with the cementitious phases led to the precipitation of calcium oxalate salts in the outer layer of the matrix. The concentration of oxalate was reduced by 65% inside the leaching medium.

  1. SimpleITK Image-Analysis Notebooks: a Collaborative Environment for Education and Reproducible Research.

    PubMed

    Yaniv, Ziv; Lowekamp, Bradley C; Johnson, Hans J; Beare, Richard

    2018-06-01

    Modern scientific endeavors increasingly require team collaborations to construct and interpret complex computational workflows. This work describes an image-analysis environment that supports the use of computational tools that facilitate reproducible research and support scientists with varying levels of software development skills. The Jupyter notebook web application is the basis of an environment that enables flexible, well-documented, and reproducible workflows via literate programming. Image-analysis software development is made accessible to scientists with varying levels of programming experience via the use of the SimpleITK toolkit, a simplified interface to the Insight Segmentation and Registration Toolkit. Additional features of the development environment include user friendly data sharing using online data repositories and a testing framework that facilitates code maintenance. SimpleITK provides a large number of examples illustrating educational and research-oriented image analysis workflows for free download from GitHub under an Apache 2.0 license: github.com/InsightSoftwareConsortium/SimpleITK-Notebooks .

  2. Water Resources Research Institute | Mississippi State University

    Science.gov Websites

    Welcome The Mississippi Water Resources Research Institute provides a statewide center of expertise with water policy issues supports state water agencies' mission with research on problems encountered in water and associated land-use and serves as a repository of knowledge for use in education

  3. Frame of Reference: Open Access Starts with You

    ERIC Educational Resources Information Center

    Goetsch, Lori A.

    2010-01-01

    Federal legislation now requires the deposit of some taxpayer-funded research in "open-access" repositories--that is, sites where scholarship and research are made freely available over the Internet. The institutions whose faculty produce the research have begun to see the benefit of open-access publication as well. From the perspective of faculty…

  4. Supporting Student Research with Semantic Technologies and Digital Archives

    ERIC Educational Resources Information Center

    Martinez-Garcia, Agustina; Corti, Louise

    2012-01-01

    This article discusses how the idea of higher education students as producers of knowledge rather than consumers can be operationalised by means of student research projects, in which processes of research archiving and analysis are enabled through the use of semantic technologies. It discusses how existing digital repository frameworks can be…

  5. Fukushima Daiichi Information Repository FY13 Status

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Smith, Curtis; Phelan, Cherie; Schwieder, Dave

    The accident at the Fukushima Daiichi nuclear power station in Japan is one of the most serious in commercial nuclear power plant operating history. Much will be learned that may be applicable to the U.S. reactor fleet, nuclear fuel cycle facilities, and supporting systems, and the international reactor fleet. For example, lessons from Fukushima Daiichi may be applied to emergency response planning, reactor operator training, accident scenario modeling, human factors engineering, radiation protection, and accident mitigation; as well as influence U.S. policies towards the nuclear fuel cycle including power generation, and spent fuel storage, reprocessing, and disposal. This document describesmore » the database used to establish a centralized information repository to store and manage the Fukushima data that has been gathered. The data is stored in a secured (password protected and encrypted) repository that is searchable and available to researchers at diverse locations.« less

  6. Proteomics data repositories: Providing a safe haven for your data and acting as a springboard for further research

    PubMed Central

    Vizcaíno, Juan Antonio; Foster, Joseph M.; Martens, Lennart

    2010-01-01

    Despite the fact that data deposition is not a generalised fact yet in the field of proteomics, several mass spectrometry (MS) based proteomics repositories are publicly available for the scientific community. The main existing resources are: the Global Proteome Machine Database (GPMDB), PeptideAtlas, the PRoteomics IDEntifications database (PRIDE), Tranche, and NCBI Peptidome. In this review the capabilities of each of these will be described, paying special attention to four key properties: data types stored, applicable data submission strategies, supported formats, and available data mining and visualization tools. Additionally, the data contents from model organisms will be enumerated for each resource. There are other valuable smaller and/or more specialized repositories but they will not be covered in this review. Finally, the concept behind the ProteomeXchange consortium, a collaborative effort among the main resources in the field, will be introduced. PMID:20615486

  7. National Database for Autism Research (NDAR): Big Data Opportunities for Health Services Research and Health Technology Assessment.

    PubMed

    Payakachat, Nalin; Tilford, J Mick; Ungar, Wendy J

    2016-02-01

    The National Database for Autism Research (NDAR) is a US National Institutes of Health (NIH)-funded research data repository created by integrating heterogeneous datasets through data sharing agreements between autism researchers and the NIH. To date, NDAR is considered the largest neuroscience and genomic data repository for autism research. In addition to biomedical data, NDAR contains a large collection of clinical and behavioral assessments and health outcomes from novel interventions. Importantly, NDAR has a global unique patient identifier that can be linked to aggregated individual-level data for hypothesis generation and testing, and for replicating research findings. As such, NDAR promotes collaboration and maximizes public investment in the original data collection. As screening and diagnostic technologies as well as interventions for children with autism are expensive, health services research (HSR) and health technology assessment (HTA) are needed to generate more evidence to facilitate implementation when warranted. This article describes NDAR and explains its value to health services researchers and decision scientists interested in autism and other mental health conditions. We provide a description of the scope and structure of NDAR and illustrate how data are likely to grow over time and become available for HSR and HTA.

  8. PSI:Biology-Materials Repository: A Biologist’s Resource for Protein Expression Plasmids

    PubMed Central

    Cormier, Catherine Y.; Park, Jin G.; Fiacco, Michael; Steel, Jason; Hunter, Preston; Kramer, Jason; Singla, Rajeev; LaBaer, Joshua

    2011-01-01

    The Protein Structure Initiative:Biology-Materials Repository (PSI:Biology-MR; MR; http://psimr.asu.edu) sequence-verifies, annotates, stores, and distributes the protein expression plasmids and vectors created by the Protein Structure Initiative (PSI). The MR has developed an informatics and sample processing pipeline that manages this process for thousands of samples per month from nearly a dozen PSI centers. DNASU (http://dnasu.asu.edu), a freely searchable database, stores the plasmid annotations, which include the full-length sequence, vector information, and associated publications for over 130,000 plasmids created by our laboratory, by the PSI and other consortia, and by individual laboratories for distribution to researchers worldwide. Each plasmid links to external resources, including the PSI Structural Biology Knowledgebase (http://sbkb.org), which facilitates cross-referencing of a particular plasmid to additional protein annotations and experimental data. To expedite and simplify plasmid requests, the MR uses an expedited material transfer agreement (EP-MTA) network, where researchers from network institutions can order and receive PSI plasmids without institutional delays. Currently over 39,000 protein expression plasmids and 78 empty vectors from the PSI are available upon request from DNASU. Overall, the MR’s repository of expression-ready plasmids, its automated pipeline, and the rapid process for receiving and distributing these plasmids more effectively allows the research community to dissect the biological function of proteins whose structures have been studied by the PSI. PMID:21360289

  9. Datasets2Tools, repository and search engine for bioinformatics datasets, tools and canned analyses

    PubMed Central

    Torre, Denis; Krawczuk, Patrycja; Jagodnik, Kathleen M.; Lachmann, Alexander; Wang, Zichen; Wang, Lily; Kuleshov, Maxim V.; Ma’ayan, Avi

    2018-01-01

    Biomedical data repositories such as the Gene Expression Omnibus (GEO) enable the search and discovery of relevant biomedical digital data objects. Similarly, resources such as OMICtools, index bioinformatics tools that can extract knowledge from these digital data objects. However, systematic access to pre-generated ‘canned’ analyses applied by bioinformatics tools to biomedical digital data objects is currently not available. Datasets2Tools is a repository indexing 31,473 canned bioinformatics analyses applied to 6,431 datasets. The Datasets2Tools repository also contains the indexing of 4,901 published bioinformatics software tools, and all the analyzed datasets. Datasets2Tools enables users to rapidly find datasets, tools, and canned analyses through an intuitive web interface, a Google Chrome extension, and an API. Furthermore, Datasets2Tools provides a platform for contributing canned analyses, datasets, and tools, as well as evaluating these digital objects according to their compliance with the findable, accessible, interoperable, and reusable (FAIR) principles. By incorporating community engagement, Datasets2Tools promotes sharing of digital resources to stimulate the extraction of knowledge from biomedical research data. Datasets2Tools is freely available from: http://amp.pharm.mssm.edu/datasets2tools. PMID:29485625

  10. Datasets2Tools, repository and search engine for bioinformatics datasets, tools and canned analyses.

    PubMed

    Torre, Denis; Krawczuk, Patrycja; Jagodnik, Kathleen M; Lachmann, Alexander; Wang, Zichen; Wang, Lily; Kuleshov, Maxim V; Ma'ayan, Avi

    2018-02-27

    Biomedical data repositories such as the Gene Expression Omnibus (GEO) enable the search and discovery of relevant biomedical digital data objects. Similarly, resources such as OMICtools, index bioinformatics tools that can extract knowledge from these digital data objects. However, systematic access to pre-generated 'canned' analyses applied by bioinformatics tools to biomedical digital data objects is currently not available. Datasets2Tools is a repository indexing 31,473 canned bioinformatics analyses applied to 6,431 datasets. The Datasets2Tools repository also contains the indexing of 4,901 published bioinformatics software tools, and all the analyzed datasets. Datasets2Tools enables users to rapidly find datasets, tools, and canned analyses through an intuitive web interface, a Google Chrome extension, and an API. Furthermore, Datasets2Tools provides a platform for contributing canned analyses, datasets, and tools, as well as evaluating these digital objects according to their compliance with the findable, accessible, interoperable, and reusable (FAIR) principles. By incorporating community engagement, Datasets2Tools promotes sharing of digital resources to stimulate the extraction of knowledge from biomedical research data. Datasets2Tools is freely available from: http://amp.pharm.mssm.edu/datasets2tools.

  11. Construction of quality-assured infant feeding process of care data repositories: definition and design (Part 1).

    PubMed

    Garcí A-de-León-Chocano, Ricardo; Sáez, Carlos; Muñoz-Soler, Verónica; Garcí A-de-León-González, Ricardo; García-Gómez, Juan M

    2015-12-01

    This is the first paper of a series of two regarding the construction of data quality (DQ) assured repositories for the reuse of information on infant feeding from birth until two years old. This first paper justifies the need for such repositories and describes the design of a process to construct them from Electronic Health Records (EHR). As a result, Part 1 proposes a computational process to obtain quality-assured datasets represented by a canonical structure extracted from raw data from multiple EHR. For this, 13 steps were defined to ensure the harmonization, standardization, completion, de-duplication, and consistency of the dataset content. Moreover, the quality of the input and output data for each of these steps is controlled according to eight DQ dimensions: predictive value, correctness, duplication, consistency, completeness, contextualization, temporal-stability and spatial-stability. The second paper of the series will describe the application of this computational process to construct the first quality-assured repository for the reuse of information on infant feeding in the perinatal period aimed at the monitoring of clinical activities and research. Copyright © 2015 Elsevier Ltd. All rights reserved.

  12. A Linked Dataset of Medical Educational Resources

    ERIC Educational Resources Information Center

    Dietze, Stefan; Taibi, Davide; Yu, Hong Qing; Dovrolis, Nikolas

    2015-01-01

    Reusable educational resources became increasingly important for enhancing learning and teaching experiences, particularly in the medical domain where resources are particularly expensive to produce. While interoperability across educational resources metadata repositories is yet limited to the heterogeneity of metadata standards and interface…

  13. Implementation and assessment of a yeast orphan gene research project; involving undergraduates in authentic research experiences and progressing our understanding of uncharacterized open reading frames

    PubMed Central

    Bowling, Bethany V.; Schultheis, Patrick J.

    2015-01-01

    Saccharomyces cerevisiae was the first eukaryotic organism to be sequenced, however little progress has been made in recent years in furthering our understanding of all open reading frames (ORFs). From October 2012 to May 2015 the number of verified ORFs has only risen from 75.31% to 78% while the number of uncharacterized ORFs have decreased from 12.8% to 11% (representing more than 700 genes still left in this category) [http://www.yeastgenome.org/genomesnapshot]. Course-based research has been shown to increase student learning while providing experience with real scientific investigation; however, implementation in large, multi-section courses presents many challenges. This study sought to test the feasibility and effectiveness of incorporating authentic research into a core genetics course with multiple instructors to increase student learning and progress our understanding of uncharacterized ORFs. We generated a module-based annotation toolkit and utilized easily accessible bioinformatics tools to predict gene function for uncharacterized ORFs within the Saccharomyces Genome Database (SGD). Students were each assigned an uncharacterized ORF which they annotated using contemporary comparative genomics methodologies including multiple sequence alignment, conserved domain identification, signal peptide prediction and cellular localization algorithms. Student learning outcomes were measured by quizzes, project reports and presentations, as well as a post-project questionnaire. Our results indicate the authentic research experience had positive impacts on student's perception of their learning and their confidence to conduct future research. Furthermore we believe that creation of an online repository and adoption and/or adaptation of this project across multiple researchers and institutions could speed the process of gene function prediction. PMID:26460164

  14. Implementation and assessment of a yeast orphan gene research project: involving undergraduates in authentic research experiences and progressing our understanding of uncharacterized open reading frames.

    PubMed

    Bowling, Bethany V; Schultheis, Patrick J; Strome, Erin D

    2016-02-01

    Saccharomyces cerevisiae was the first eukaryotic organism to be sequenced; however, little progress has been made in recent years in furthering our understanding of all open reading frames (ORFs). From October 2012 to May 2015 the number of verified ORFs had only risen from 75.31% to 78%, while the number of uncharacterized ORFs had decreased from 12.8% to 11% (representing > 700 genes still left in this category; http://www.yeastgenome.org/genomesnapshot). Course-based research has been shown to increase student learning while providing experience with real scientific investigation; however, implementation in large, multi-section courses presents many challenges. This study sought to test the feasibility and effectiveness of incorporating authentic research into a core genetics course, with multiple instructors, to increase student learning and progress our understanding of uncharacterized ORFs. We generated a module-based annotation toolkit and utilized easily accessible bioinformatics tools to predict gene function for uncharacterized ORFs within the Saccharomyces Genome Database (SGD). Students were each assigned an uncharacterized ORF, which they annotated using contemporary comparative genomics methodologies, including multiple sequence alignment, conserved domain identification, signal peptide prediction and cellular localization algorithms. Student learning outcomes were measured by quizzes, project reports and presentations, as well as a post-project questionnaire. Our results indicate that the authentic research experience had positive impacts on students' perception of their learning and their confidence to conduct future research. Furthermore, we believe that creation of an online repository and adoption and/or adaptation of this project across multiple researchers and institutions could speed the process of gene function prediction. Copyright © 2015 John Wiley & Sons, Ltd.

  15. Repository-Based Software Engineering (RBSE) program

    NASA Technical Reports Server (NTRS)

    1992-01-01

    Support of a software engineering program was provided in the following areas: client/customer liaison; research representation/outreach; and program support management. Additionally, a list of deliverables is presented.

  16. A Framework to Support the Sharing and Reuse of Computable Phenotype Definitions Across Health Care Delivery and Clinical Research Applications.

    PubMed

    Richesson, Rachel L; Smerek, Michelle M; Blake Cameron, C

    2016-01-01

    The ability to reproducibly identify clinically equivalent patient populations is critical to the vision of learning health care systems that implement and evaluate evidence-based treatments. The use of common or semantically equivalent phenotype definitions across research and health care use cases will support this aim. Currently, there is no single consolidated repository for computable phenotype definitions, making it difficult to find all definitions that already exist, and also hindering the sharing of definitions between user groups. Drawing from our experience in an academic medical center that supports a number of multisite research projects and quality improvement studies, we articulate a framework that will support the sharing of phenotype definitions across research and health care use cases, and highlight gaps and areas that need attention and collaborative solutions. An infrastructure for re-using computable phenotype definitions and sharing experience across health care delivery and clinical research applications includes: access to a collection of existing phenotype definitions, information to evaluate their appropriateness for particular applications, a knowledge base of implementation guidance, supporting tools that are user-friendly and intuitive, and a willingness to use them. We encourage prospective researchers and health administrators to re-use existing EHR-based condition definitions where appropriate and share their results with others to support a national culture of learning health care. There are a number of federally funded resources to support these activities, and research sponsors should encourage their use.

  17. A Framework to Support the Sharing and Reuse of Computable Phenotype Definitions Across Health Care Delivery and Clinical Research Applications

    PubMed Central

    Richesson, Rachel L.; Smerek, Michelle M.; Blake Cameron, C.

    2016-01-01

    Introduction: The ability to reproducibly identify clinically equivalent patient populations is critical to the vision of learning health care systems that implement and evaluate evidence-based treatments. The use of common or semantically equivalent phenotype definitions across research and health care use cases will support this aim. Currently, there is no single consolidated repository for computable phenotype definitions, making it difficult to find all definitions that already exist, and also hindering the sharing of definitions between user groups. Method: Drawing from our experience in an academic medical center that supports a number of multisite research projects and quality improvement studies, we articulate a framework that will support the sharing of phenotype definitions across research and health care use cases, and highlight gaps and areas that need attention and collaborative solutions. Framework: An infrastructure for re-using computable phenotype definitions and sharing experience across health care delivery and clinical research applications includes: access to a collection of existing phenotype definitions, information to evaluate their appropriateness for particular applications, a knowledge base of implementation guidance, supporting tools that are user-friendly and intuitive, and a willingness to use them. Next Steps: We encourage prospective researchers and health administrators to re-use existing EHR-based condition definitions where appropriate and share their results with others to support a national culture of learning health care. There are a number of federally funded resources to support these activities, and research sponsors should encourage their use. PMID:27563686

  18. Federated queries of clinical data repositories: Scaling to a national network.

    PubMed

    Weber, Griffin M

    2015-06-01

    Federated networks of clinical research data repositories are rapidly growing in size from a handful of sites to true national networks with more than 100 hospitals. This study creates a conceptual framework for predicting how various properties of these systems will scale as they continue to expand. Starting with actual data from Harvard's four-site Shared Health Research Information Network (SHRINE), the framework is used to imagine a future 4000 site network, representing the majority of hospitals in the United States. From this it becomes clear that several common assumptions of small networks fail to scale to a national level, such as all sites being online at all times or containing data from the same date range. On the other hand, a large network enables researchers to select subsets of sites that are most appropriate for particular research questions. Developers of federated clinical data networks should be aware of how the properties of these networks change at different scales and design their software accordingly. Copyright © 2015 Elsevier Inc. All rights reserved.

  19. A Unique Digital Electrocardiographic Repository for the Development of Quantitative Electrocardiography and Cardiac Safety: The Telemetric and Holter ECG Warehouse (THEW)

    PubMed Central

    Couderc, Jean-Philippe

    2010-01-01

    The sharing of scientific data reinforces open scientific inquiry; it encourages diversity of analysis and opinion while promoting new research and facilitating the education of next generations of scientists. In this article, we present an initiative for the development of a repository containing continuous electrocardiographic information and their associated clinical information. This information is shared with the worldwide scientific community in order to improve quantitative electrocardiology and cardiac safety. First, we present the objectives of the initiative and its mission. Then, we describe the resources available in this initiative following three components: data, expertise and tools. The Data available in the Telemetric and Holter ECG Warehouse (THEW) includes continuous ECG signals and associated clinical information. The initiative attracted various academic and private partners whom expertise covers a large list of research arenas related to quantitative electrocardiography; their contribution to the THEW promotes cross-fertilization of scientific knowledge, resources, and ideas that will advance the field of quantitative electrocardiography. Finally, the tools of the THEW include software and servers to access and review the data available in the repository. To conclude, the THEW is an initiative developed to benefit the scientific community and to advance the field of quantitative electrocardiography and cardiac safety. It is a new repository designed to complement the existing ones such as Physionet, the AHA-BIH Arrhythmia Database, and the CSE database. The THEW hosts unique datasets from clinical trials and drug safety studies that, so far, were not available to the worldwide scientific community. PMID:20863512

  20. Description of the datasets for the experiments in the paper "solving the multi-vehicle multi-covering tour problem".

    PubMed

    Pham, Tuan Anh; Hà, Minh Hoàng; Nguyen, Xuan Hoai

    2018-06-01

    This data article contains data related to the research article entitled, "Solving the multi-vehicle multi-covering tour problem" (Pham et al., 2017) [4]. All data of this article was generated from instances kroA100, kroB100, kroC100, kroD100, kroA200, and kroB200 from TSPLIB. It can be downloaded from public repository. This data can be used as benchmarks for the covering tour problem (CTP) variants, such as m -CTP- p , m -CTP, mm -CTP- p , mm -CTP, mm -CTP-o, mm -CTP-wo. We tested our algorithm on these data and results are shown in Pham et al. (2017) [4].

  1. Interfaces to PeptideAtlas: a case study of standard data access systems

    PubMed Central

    Handcock, Jeremy; Robinson, Thomas; Deutsch, Eric W.; Boyle, John

    2012-01-01

    Access to public data sets is important to the scientific community as a resource to develop new experiments or validate new data. Projects such as the PeptideAtlas, Ensembl and The Cancer Genome Atlas (TCGA) offer both access to public data and a repository to share their own data. Access to these data sets is often provided through a web page form and a web service API. Access technologies based on web protocols (e.g. http) have been in use for over a decade and are widely adopted across the industry for a variety of functions (e.g. search, commercial transactions, and social media). Each architecture adapts these technologies to provide users with tools to access and share data. Both commonly used web service technologies (e.g. REST and SOAP), and custom-built solutions over HTTP are utilized in providing access to research data. Providing multiple access points ensures that the community can access the data in the simplest and most effective manner for their particular needs. This article examines three common access mechanisms for web accessible data: BioMart, caBIG, and Google Data Sources. These are illustrated by implementing each over the PeptideAtlas repository and reviewed for their suitability based on specific usages common to research. BioMart, Google Data Sources, and caBIG are each suitable for certain uses. The tradeoffs made in the development of the technology are dependent on the uses each was designed for (e.g. security versus speed). This means that an understanding of specific requirements and tradeoffs is necessary before selecting the access technology. PMID:22941959

  2. Utility of the Department of Defense Serum Repository in Assessing Deployment Exposure.

    PubMed

    Lushniak, Boris; Mallon, Col Timothy M; Gaydos, Joel C; Smith, David J

    2016-08-01

    This paper describes why the research project was conducted in terms of demonstrating the utility of the Department of Defense Serum Repository in addressing deployment environmental exposures. The history deployment exposure surveillance was reviewed and the rationale for developing validated biomarkers that were detected in sera in postdeployment samples and compared with nondeployed controls was described. The goal was to find validated biomarkers that are associated with both exposures and health outcomes. The articles in this supplement described novel serum biomarkers that were found to be associated with deployment exposures and weakly associated with some health outcomes. Future research must continue to validate the use of serum biomarkers when operational contingencies prevent the gold standard collection of real-time breathing zone samples in deployed service members.

  3. Simulator sickness research program at NASA-Ames Research Center

    NASA Technical Reports Server (NTRS)

    Mccauley, Michael E.; Cook, Anthony M.

    1987-01-01

    The simulator sickness syndrome is receiving increased attention in the simulation community. NASA-Ames Research Center has initiated a program to facilitate the exchange of information on this topic among the tri-services and other interested government organizations. The program objectives are to identify priority research issues, promote efficient research strategies, serve as a repository of information, and disseminate information to simulator users.

  4. Donor human milk bank data collection in north america: an assessment of current status and future needs.

    PubMed

    Brownell, Elizabeth A; Lussier, Mary M; Herson, Victor C; Hagadorn, James I; Marinelli, Kathleen A

    2014-02-01

    The Human Milk Banking Association of North America (HMBANA) is a nonprofit association that standardizes and facilitates the establishment and operation of donor human milk (DHM) banks in North America. Each HMBANA milk bank in the network collects data on the DHM it receives and distributes, but a centralized data repository does not yet exist. In 2010, the Food and Drug Administration recognized the need to collect and disseminate systematic, standardized DHM bank data and suggested that HMBANA develop a DHM data repository. This study aimed to describe data currently collected by HMBANA DHM banks and evaluate feasibility and interest in participating in a centralized data repository. We conducted phone interviews with individuals in different HMBANA milk banks and summarized descriptive statistics. Eight of 13 (61.5%) sites consented to participate. All respondents collected donor demographics, and half (50%; n = 4) rescreened donors after 6 months of continued donation. The definition of preterm milk varied between DHM banks (≤ 32 to ≤ 40 weeks). The specific computer program used to house the data also differed. Half (50%; n = 4) indicated that they would consider participation in a centralized repository. Without standardized data across all HMBANA sites, the creation of a centralized data repository is not yet feasible. Lack of standardization and transparency may deter implementation of donor milk programs in the neonatal intensive care unit setting and hinder benchmarking, research, and quality improvement initiatives.

  5. [Self-archiving of biomedical papers in open access repositories].

    PubMed

    Abad-García, M Francisca; Melero, Remedios; Abadal, Ernest; González-Teruel, Aurora

    2010-04-01

    Open-access literature is digital, online, free of charge, and free of most copyright and licensing restrictions. Self-archiving or deposit of scholarly outputs in institutional repositories (open-access green route) is increasingly present in the activities of the scientific community. Besides the benefits of open access for visibility and dissemination of science, it is increasingly more often required by funding agencies to deposit papers and any other type of documents in repositories. In the biomedical environment this is even more relevant by the impact scientific literature can have on public health. However, to make self-archiving feasible, authors should be aware of its meaning and the terms in which they are allowed to archive their works. In that sense, there are some tools like Sherpa/RoMEO or DULCINEA (both directories of copyright licences of scientific journals at different levels) to find out what rights are retained by authors when they publish a paper and if they allow to implement self-archiving. PubMed Central and its British and Canadian counterparts are the main thematic repositories for biomedical fields. In our country there is none of similar nature, but most of the universities and CSIC, have already created their own institutional repositories. The increase in visibility of research results and their impact on a greater and earlier citation is one of the most frequently advance of open access, but removal of economic barriers to access to information is also a benefit to break borders between groups.

  6. A strategy to establish Food Safety Model Repositories.

    PubMed

    Plaza-Rodríguez, C; Thoens, C; Falenski, A; Weiser, A A; Appel, B; Kaesbohrer, A; Filter, M

    2015-07-02

    Transferring the knowledge of predictive microbiology into real world food manufacturing applications is still a major challenge for the whole food safety modelling community. To facilitate this process, a strategy for creating open, community driven and web-based predictive microbial model repositories is proposed. These collaborative model resources could significantly improve the transfer of knowledge from research into commercial and governmental applications and also increase efficiency, transparency and usability of predictive models. To demonstrate the feasibility, predictive models of Salmonella in beef previously published in the scientific literature were re-implemented using an open source software tool called PMM-Lab. The models were made publicly available in a Food Safety Model Repository within the OpenML for Predictive Modelling in Food community project. Three different approaches were used to create new models in the model repositories: (1) all information relevant for model re-implementation is available in a scientific publication, (2) model parameters can be imported from tabular parameter collections and (3) models have to be generated from experimental data or primary model parameters. All three approaches were demonstrated in the paper. The sample Food Safety Model Repository is available via: http://sourceforge.net/projects/microbialmodelingexchange/files/models and the PMM-Lab software can be downloaded from http://sourceforge.net/projects/pmmlab/. This work also illustrates that a standardized information exchange format for predictive microbial models, as the key component of this strategy, could be established by adoption of resources from the Systems Biology domain. Copyright © 2015. Published by Elsevier B.V.

  7. Restricting access to publications from funded research: ethical issues and solutions.

    PubMed

    Manikandan, S; Vani, N Isai

    2010-01-01

    India is becoming one of the hubs of clinical research. Commensurate with these advances, the government funding for biomedical research in thrust areas is also increasing. The Indian Council of Medical Research (ICMR), Department of Biotechnology (DBT), Department of Science and Technology (DST) are some of the government organizations which provide financial support for various research projects. The results of the funded research projects are published in various international journals. Most of these journals have an access to paid subscribers only. Hence it is unethical to use the research grants from government (people's money) and not allow the scientific community free access to the results of the study. To tackle such issues, these agencies should sign the Berlin declaration and create open access repositories. A public access policy should be formulated and listed in JULIET. The funding bodies in India should also join Pubmed Central (PMC) to form PMC India so that every investigator who has received grants would submit the full text of the paper published from his study and these can be made freely accessible to everyone. Universities and research institutions should also develop institutional open access repositories. The public access policy has definitive advantages and should be implemented.

  8. Generic Argillite/Shale Disposal Reference Case

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zheng, Liange; Colon, Carlos Jové; Bianchi, Marco

    Radioactive waste disposal in a deep subsurface repository hosted in clay/shale/argillite is a subject of widespread interest given the desirable isolation properties, geochemically reduced conditions, and widespread geologic occurrence of this rock type (Hansen 2010; Bianchi et al. 2013). Bianchi et al. (2013) provides a description of diffusion in a clay-hosted repository based on single-phase flow and full saturation using parametric data from documented studies in Europe (e.g., ANDRA 2005). The predominance of diffusive transport and sorption phenomena in this clay media are key attributes to impede radionuclide mobility making clay rock formations target sites for disposal of high-level radioactivemore » waste. The reports by Hansen et al. (2010) and those from numerous studies in clay-hosted underground research laboratories (URLs) in Belgium, France and Switzerland outline the extensive scientific knowledge obtained to assess long-term clay/shale/argillite repository isolation performance of nuclear waste. In the past several years under the UFDC, various kinds of models have been developed for argillite repository to demonstrate the model capability, understand the spatial and temporal alteration of the repository, and evaluate different scenarios. These models include the coupled Thermal-Hydrological-Mechanical (THM) and Thermal-Hydrological-Mechanical-Chemical (THMC) models (e.g. Liu et al. 2013; Rutqvist et al. 2014a, Zheng et al. 2014a) that focus on THMC processes in the Engineered Barrier System (EBS) bentonite and argillite host hock, the large scale hydrogeologic model (Bianchi et al. 2014) that investigates the hydraulic connection between an emplacement drift and surrounding hydrogeological units, and Disposal Systems Evaluation Framework (DSEF) models (Greenberg et al. 2013) that evaluate thermal evolution in the host rock approximated as a thermal conduction process to facilitate the analysis of design options. However, the assumptions and the properties (parameters) used in these models are different, which not only make inter-model comparisons difficult, but also compromise the applicability of the lessons learned from one model to another model. The establishment of a reference case would therefore be helpful to set up a baseline for model development. A generic salt repository reference case was developed in Freeze et al. (2013) and the generic argillite repository reference case is presented in this report. The definition of a reference case requires the characterization of the waste inventory, waste form, waste package, repository layout, EBS backfill, host rock, and biosphere. This report mainly documents the processes in EBS bentonite and host rock that are potentially important for performance assessment and properties that are needed to describe these processes, with brief description other components such as waste inventory, waste form, waste package, repository layout, aquifer, and biosphere. A thorough description of the generic argillite repository reference case will be given in Jové Colon et al. (2014).« less

  9. 10 CFR 60.140 - General requirements.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... and it will continue until permanent closure. (c) The program shall include in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to accomplish the objective as... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN GEOLOGIC REPOSITORIES...

  10. 10 CFR 60.140 - General requirements.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... and it will continue until permanent closure. (c) The program shall include in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to accomplish the objective as... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN GEOLOGIC REPOSITORIES...

  11. 10 CFR 60.140 - General requirements.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... and it will continue until permanent closure. (c) The program shall include in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to accomplish the objective as... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN GEOLOGIC REPOSITORIES...

  12. 10 CFR 63.131 - General requirements.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN A GEOLOGIC REPOSITORY AT YUCCA... conditions encountered and changes in those conditions during construction and waste emplacement operations...

  13. 10 CFR 60.140 - General requirements.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... and it will continue until permanent closure. (c) The program shall include in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to accomplish the objective as... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN GEOLOGIC REPOSITORIES...

  14. 10 CFR 63.131 - General requirements.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN A GEOLOGIC REPOSITORY AT YUCCA... conditions encountered and changes in those conditions during construction and waste emplacement operations...

  15. 10 CFR 63.131 - General requirements.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN A GEOLOGIC REPOSITORY AT YUCCA... conditions encountered and changes in those conditions during construction and waste emplacement operations...

  16. 10 CFR 63.131 - General requirements.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN A GEOLOGIC REPOSITORY AT YUCCA... conditions encountered and changes in those conditions during construction and waste emplacement operations...

  17. 10 CFR 60.140 - General requirements.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... and it will continue until permanent closure. (c) The program shall include in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to accomplish the objective as... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN GEOLOGIC REPOSITORIES...

  18. 10 CFR 63.131 - General requirements.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... in situ monitoring, laboratory and field testing, and in situ experiments, as may be appropriate to... REGULATORY COMMISSION (CONTINUED) DISPOSAL OF HIGH-LEVEL RADIOACTIVE WASTES IN A GEOLOGIC REPOSITORY AT YUCCA... conditions encountered and changes in those conditions during construction and waste emplacement operations...

  19. 75 FR 71133 - National Institute of Mental Health; Notice of Closed Meeting

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-11-22

    ... Emphasis Panel; Competitive Revision for Stem Cell Repository Relevant to Mental Disorders. Date: December... Domestic Assistance Program Nos. 93.242, Mental Health Research Grants; 93.281, Scientist Development Award, Scientist Development Award for Clinicians, and Research Scientist Award; 93.282, Mental Health National...

  20. Data Publication: A Partnership between Scientists, Data Managers and Librarians

    NASA Astrophysics Data System (ADS)

    Raymond, L.; Chandler, C.; Lowry, R.; Urban, E.; Moncoiffe, G.; Pissierssens, P.; Norton, C.; Miller, H.

    2012-04-01

    Current literature on the topic of data publication suggests that success is best achieved when there is a partnership between scientists, data managers, and librarians. The Marine Biological Laboratory/Woods Hole Oceanographic Institution (MBLWHOI) Library and the Biological and Chemical Oceanography Data Management Office (BCO-DMO) have developed tools and processes to automate the ingestion of metadata from BCO-DMO for deposit with datasets into the Institutional Repository (IR) Woods Hole Open Access Server (WHOAS). The system also incorporates functionality for BCO-DMO to request a Digital Object Identifier (DOI) from the Library. This partnership allows the Library to work with a trusted data repository to ensure high quality data while the data repository utilizes library services and is assured of a permanent archive of the copy of the data extracted from the repository database. The assignment of persistent identifiers enables accurate data citation. The Library can assign a DOI to appropriate datasets deposited in WHOAS. A primary activity is working with authors to deposit datasets associated with published articles. The DOI would ideally be assigned before submission and be included in the published paper so readers can link directly to the dataset, but DOIs are also being assigned to datasets related to articles after publication. WHOAS metadata records link the article to the datasets and the datasets to the article. The assignment of DOIs has enabled another important collaboration with Elsevier, publisher of educational and professional science journals. Elsevier can now link from articles in the Science Direct database to the datasets available from WHOAS that are related to that article. The data associated with the article are freely available from WHOAS and accompanied by a Dublin Core metadata record. In addition, the Library has worked with researchers to deposit datasets in WHOAS that are not appropriate for national, international, or domain specific data repositories. These datasets currently include audio, text and image files. This research is being conducted by a team of librarians, data managers and scientists that are collaborating with representatives from the Scientific Committee on Oceanic Research (SCOR) and the International Oceanographic Data and Information Exchange (IODE) of the Intergovernmental Oceanographic Commission (IOC). The goal is to identify best practices for tracking data provenance and clearly attributing credit to data collectors/providers.

  1. Transportation needs assessment: Emergency response section

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    NONE

    The transportation impacts of moving high level nuclear waste (HLNW) to a repository at Yucca Mountain in Nevada are of concern to the residents of the State as well as to the residents of other states through which the nuclear wastes might be transported. The projected volume of the waste suggests that shipments will occur on a daily basis for some period of time. This will increase the risk of accidents, including a catastrophic incident. Furthermore, as the likelihood of repository construction and operation and waste shipments increase, so will the attention given by the national media. This document ismore » not to be construed as a willingness to accept the HLNW repository on the part of the State. Rather it is an initial step in ensuring that the safety and well-being of Nevada residents and visitors and the State`s economy will be adequately addressed in federal decision-making pertaining to the transportation of HLNW into and across Nevada for disposal in the proposed repository. The Preferred Transportation System Needs Assessment identifies critical system design elements and technical and social issues that must be considered in conducting a comprehensive transportation impact analysis. Development of the needs assessment and the impact analysis is especially complex because of the absence of information and experience with shipping HLNW and because of the ``low probability, high consequence`` aspect of the transportation risk.« less

  2. Safeguarding structural data repositories against bad apples

    PubMed Central

    Minor, Wladek; Dauter, Zbigniew; Helliwell, John R.; Jaskolski, Mariusz; Wlodawer, Alexander

    2016-01-01

    Structural biology research generates large amounts of data, some deposited in public databases/repositories, but a substantial remainder never becoming available to the scientific community. Additionally, some of the deposited data contain less or more serious errors that may bias the results of data mining. Thorough analysis and discussion of these problems is needed in order to ameliorate this situation. This note is an attempt to propose some solutions and encourage both further discussion and action on the part of the relevant organizations, in particular the Protein Data Bank and various bodies of the International Union of Crystallography. PMID:26840827

  3. Grid Application Meta-Repository System: Repository Interconnectivity and Cross-domain Application Usage in Distributed Computing Environments

    NASA Astrophysics Data System (ADS)

    Tudose, Alexandru; Terstyansky, Gabor; Kacsuk, Peter; Winter, Stephen

    Grid Application Repositories vary greatly in terms of access interface, security system, implementation technology, communication protocols and repository model. This diversity has become a significant limitation in terms of interoperability and inter-repository access. This paper presents the Grid Application Meta-Repository System (GAMRS) as a solution that offers better options for the management of Grid applications. GAMRS proposes a generic repository architecture, which allows any Grid Application Repository (GAR) to be connected to the system independent of their underlying technology. It also presents applications in a uniform manner and makes applications from all connected repositories visible to web search engines, OGSI/WSRF Grid Services and other OAI (Open Archive Initiative)-compliant repositories. GAMRS can also function as a repository in its own right and can store applications under a new repository model. With the help of this model, applications can be presented as embedded in virtual machines (VM) and therefore they can be run in their native environments and can easily be deployed on virtualized infrastructures allowing interoperability with new generation technologies such as cloud computing, application-on-demand, automatic service/application deployments and automatic VM generation.

  4. Leveraging Radioactive Waste Disposal at WIPP for Science

    NASA Astrophysics Data System (ADS)

    Rempe, N. T.

    2008-12-01

    Salt mines are radiologically much quieter than other underground environments because of ultra-low concentrations of natural radionuclides (U, Th, and K) in the host rock; therefore, the Waste Isolation Pilot Plant (WIPP), a government-owned, 655m deep geologic repository that disposes of radioactive waste in thick salt near Carlsbad, New Mexico, has for the last 15 years hosted highly radiation-sensitive experiments. Incidentally, Nature started her own low background experiment 250ma ago, preserving viable bacteria, cellulose, and DNA in WIPP salt. The Department of Energy continues to make areas of the WIPP underground available for experiments, freely offering its infrastructure and access to this unique environment. Even before WIPP started disposing of waste in 1999, the Room-Q alcove (25m x 10m x 4m) housed a succession of small experiments. They included development and calibration of neutral-current detectors by Los Alamos National Laboratory (LANL) for the Sudbury Neutrino Observatory, a proof-of-concept by Ohio State University of a flavor-sensitive neutrino detector for supernovae, and research by LANL on small solid- state dark matter detectors. Two currently active experiments support the search for neutrino-less double beta decay as a tool to better define the nature and mass of the neutrino. That these delicate experiments are conducted in close vicinity to, but not at all affected by, megacuries of radioactive waste reinforces the safety argument for the repository. Since 2003, the Majorana collaboration is developing and testing various detector designs inside a custom- built clean room in the Room-Q alcove. Already low natural background readings are reduced further by segmenting the germanium detectors, which spatially and temporally discriminates background radiation. The collaboration also demonstrated safe copper electro-forming underground, which minimizes cosmogenic background in detector assemblies. The largest currently used experimental space (100m x 10m x 6m) is the North Experimental Area (NExA). There, Enriched Xenon Observatory (EXO) collaborators have since mid-2007 been assembling and outfitting six modules and associated structures that were pre-assembled at Stanford University, then dismantled, and shipped to WIPP. Transporting the modules underground presented several interesting challenges, all of which were overcome. Access through increasingly cleaner joined modules leads to the class-100 clean room detector module. Inside, a time projection chamber (TPC) contains 200kg liquid Xe- 136 (the largest non-defense related stockpile of an enriched isotope ever assembled for research). After the experiment starts in early 2009, it is expected to run for 3-5 years. University of Pennsylvania researchers recently sampled WIPP salt to attempt measuring stable Ne-22, resulting from the interaction of cosmogenic muons with Na-23 and preserved in the halite lattice, to determine variations in the cosmic-radiation flux. They in turn could reveal the history of nearby supernovae. University of Chicago/Fermilab researchers evaluate whether to install a superheated-fluid bubble-chamber to search for weakly interacting massive particles (WIMPs). A helium-filled solar neutrino TPC, dark matter and neutron detectors, and proton-decay and supernova-neutrino detectors are other projects that were and are under discussion. Rounding out the spectrum of possibilities are experiments to investigate the effects of long-term ultra-low-dose radiation on cell cultures and laboratory animals to verify or falsify the linear, no- threshold hypothesis. WIPP welcomes additional proposals and projects.

  5. Overview of FEED, the feeding experiments end-user database.

    PubMed

    Wall, Christine E; Vinyard, Christopher J; Williams, Susan H; Gapeyev, Vladimir; Liu, Xianhua; Lapp, Hilmar; German, Rebecca Z

    2011-08-01

    The Feeding Experiments End-user Database (FEED) is a research tool developed by the Mammalian Feeding Working Group at the National Evolutionary Synthesis Center that permits synthetic, evolutionary analyses of the physiology of mammalian feeding. The tasks of the Working Group are to compile physiologic data sets into a uniform digital format stored at a central source, develop a standardized terminology for describing and organizing the data, and carry out a set of novel analyses using FEED. FEED contains raw physiologic data linked to extensive metadata. It serves as an archive for a large number of existing data sets and a repository for future data sets. The metadata are stored as text and images that describe experimental protocols, research subjects, and anatomical information. The metadata incorporate controlled vocabularies to allow consistent use of the terms used to describe and organize the physiologic data. The planned analyses address long-standing questions concerning the phylogenetic distribution of phenotypes involving muscle anatomy and feeding physiology among mammals, the presence and nature of motor pattern conservation in the mammalian feeding muscles, and the extent to which suckling constrains the evolution of feeding behavior in adult mammals. We expect FEED to be a growing digital archive that will facilitate new research into understanding the evolution of feeding anatomy.

  6. Reproducibility in Computational Neuroscience Models and Simulations

    PubMed Central

    McDougal, Robert A.; Bulanova, Anna S.; Lytton, William W.

    2016-01-01

    Objective Like all scientific research, computational neuroscience research must be reproducible. Big data science, including simulation research, cannot depend exclusively on journal articles as the method to provide the sharing and transparency required for reproducibility. Methods Ensuring model reproducibility requires the use of multiple standard software practices and tools, including version control, strong commenting and documentation, and code modularity. Results Building on these standard practices, model sharing sites and tools have been developed that fit into several categories: 1. standardized neural simulators, 2. shared computational resources, 3. declarative model descriptors, ontologies and standardized annotations; 4. model sharing repositories and sharing standards. Conclusion A number of complementary innovations have been proposed to enhance sharing, transparency and reproducibility. The individual user can be encouraged to make use of version control, commenting, documentation and modularity in development of models. The community can help by requiring model sharing as a condition of publication and funding. Significance Model management will become increasingly important as multiscale models become larger, more detailed and correspondingly more difficult to manage by any single investigator or single laboratory. Additional big data management complexity will come as the models become more useful in interpreting experiments, thus increasing the need to ensure clear alignment between modeling data, both parameters and results, and experiment. PMID:27046845

  7. The Open Data Repositorys Data Publisher

    NASA Technical Reports Server (NTRS)

    Stone, N.; Lafuente, B.; Downs, R. T.; Blake, D.; Bristow, T.; Fonda, M.; Pires, A.

    2015-01-01

    Data management and data publication are becoming increasingly important components of researcher's workflows. The complexity of managing data, publishing data online, and archiving data has not decreased significantly even as computing access and power has greatly increased. The Open Data Repository's Data Publisher software strives to make data archiving, management, and publication a standard part of a researcher's workflow using simple, web-based tools and commodity server hardware. The publication engine allows for uploading, searching, and display of data with graphing capabilities and downloadable files. Access is controlled through a robust permissions system that can control publication at the field level and can be granted to the general public or protected so that only registered users at various permission levels receive access. Data Publisher also allows researchers to subscribe to meta-data standards through a plugin system, embargo data publication at their discretion, and collaborate with other researchers through various levels of data sharing. As the software matures, semantic data standards will be implemented to facilitate machine reading of data and each database will provide a REST application programming interface for programmatic access. Additionally, a citation system will allow snapshots of any data set to be archived and cited for publication while the data itself can remain living and continuously evolve beyond the snapshot date. The software runs on a traditional LAMP (Linux, Apache, MySQL, PHP) server and is available on GitHub (http://github.com/opendatarepository) under a GPLv2 open source license. The goal of the Open Data Repository is to lower the cost and training barrier to entry so that any researcher can easily publish their data and ensure it is archived for posterity.

  8. Digital Rocks Portal: a sustainable platform for imaged dataset sharing, translation and automated analysis

    NASA Astrophysics Data System (ADS)

    Prodanovic, M.; Esteva, M.; Hanlon, M.; Nanda, G.; Agarwal, P.

    2015-12-01

    Recent advances in imaging have provided a wealth of 3D datasets that reveal pore space microstructure (nm to cm length scale) and allow investigation of nonlinear flow and mechanical phenomena from first principles using numerical approaches. This framework has popularly been called "digital rock physics". Researchers, however, have trouble storing and sharing the datasets both due to their size and the lack of standardized image types and associated metadata for volumetric datasets. This impedes scientific cross-validation of the numerical approaches that characterize large scale porous media properties, as well as development of multiscale approaches required for correct upscaling. A single research group typically specializes in an imaging modality and/or related modeling on a single length scale, and lack of data-sharing infrastructure makes it difficult to integrate different length scales. We developed a sustainable, open and easy-to-use repository called the Digital Rocks Portal, that (1) organizes images and related experimental measurements of different porous materials, (2) improves access to them for a wider community of geosciences or engineering researchers not necessarily trained in computer science or data analysis. Once widely accepter, the repository will jumpstart productivity and enable scientific inquiry and engineering decisions founded on a data-driven basis. This is the first repository of its kind. We show initial results on incorporating essential software tools and pipelines that make it easier for researchers to store and reuse data, and for educators to quickly visualize and illustrate concepts to a wide audience. For data sustainability and continuous access, the portal is implemented within the reliable, 24/7 maintained High Performance Computing Infrastructure supported by the Texas Advanced Computing Center (TACC) at the University of Texas at Austin. Long-term storage is provided through the University of Texas System Research Cyber-infrastructure initiative.

  9. The Geodetic Seamless Archive Centers Service Layer: A System Architecture for Federating Geodesy Data Repositories

    NASA Astrophysics Data System (ADS)

    McWhirter, J.; Boler, F. M.; Bock, Y.; Jamason, P.; Squibb, M. B.; Noll, C. E.; Blewitt, G.; Kreemer, C. W.

    2010-12-01

    Three geodesy Archive Centers, Scripps Orbit and Permanent Array Center (SOPAC), NASA's Crustal Dynamics Data Information System (CDDIS) and UNAVCO are engaged in a joint effort to define and develop a common Web Service Application Programming Interface (API) for accessing geodetic data holdings. This effort is funded by the NASA ROSES ACCESS Program to modernize the original GPS Seamless Archive Centers (GSAC) technology which was developed in the 1990s. A new web service interface, the GSAC-WS, is being developed to provide uniform and expanded mechanisms through which users can access our data repositories. In total, our respective archives hold tens of millions of files and contain a rich collection of site/station metadata. Though we serve similar user communities, we currently provide a range of different access methods, query services and metadata formats. This leads to a lack of consistency in the userís experience and a duplication of engineering efforts. The GSAC-WS API and its reference implementation in an underlying Java-based GSAC Service Layer (GSL) supports metadata and data queries into site/station oriented data archives. The general nature of this API makes it applicable to a broad range of data systems. The overall goals of this project include providing consistent and rich query interfaces for end users and client programs, the development of enabling technology to facilitate third party repositories in developing these web service capabilities and to enable the ability to perform data queries across a collection of federated GSAC-WS enabled repositories. A fundamental challenge faced in this project is to provide a common suite of query services across a heterogeneous collection of data yet enabling each repository to expose their specific metadata holdings. To address this challenge we are developing a "capabilities" based service where a repository can describe its specific query and metadata capabilities. Furthermore, the architecture of the GSL is based on a model-view paradigm that decouples the underlying data model semantics from particular representations of the data model. This will allow for the GSAC-WS enabled repositories to evolve their service offerings to incorporate new metadata definition formats (e.g., ISO-19115, FGDC, JSON, etc.) and new techniques for accessing their holdings. Building on the core GSAC-WS implementations the project is also developing a federated/distributed query service. This service will seamlessly integrate with the GSAC Service Layer and will support data and metadata queries across a collection of federated GSAC repositories.

  10. Proposed BioRepository platform solution for the ALS research community.

    PubMed

    Sherman, Alex; Bowser, Robert; Grasso, Daniela; Power, Breen; Milligan, Carol; Jaffa, Matthew; Cudkowicz, Merit

    2011-01-01

    ALS is a rare disorder whose cause and pathogenesis is largely unknown ( 1 ). There is a recognized need to develop biomarkers for ALS to better understand the disease, expedite diagnosis and to facilitate therapy development. Collaboration is essential to obtain a sufficient number of samples to allow statistically meaningful studies. The availability of high quality biological specimens for research purposes requires the development of standardized methods for collection, long-term storage, retrieval and distribution of specimens. The value of biological samples to scientists and clinicians correlates with the completeness and relevance of phenotypical and clinical information associated with the samples ( 2 , 3 ). While developing a secure Web-based system to manage an inventory of multi-site BioRepositories, algorithms were implemented to facilitate ad hoc parametric searches across heterogeneous data sources that contain data from clinical trials and research studies. A flexible schema for a barcode label was introduced to allow association of samples to these data. The ALSBank™ BioRepository platform solution for managing biological samples and associated data is currently deployed by the Northeast ALS Consortium (NEALS). The NEALS Consortium and the Massachusetts General Hospital (MGH) Neurology Clinical Trials Unit (NCTU) support a network of multiple BioBanks, thus allowing researchers to take advantage of a larger specimen collection than they might have at an individual institution. Standard operating procedures are utilized at all collection sites to promote common practices for biological sample integrity, quality control and associated clinical data. Utilizing this platform, we have created one of the largest virtual collections of ALS-related specimens available to investigators studying ALS.

  11. DR Argillite Disposal R&D at LBNL

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zheng, Liange; Kim, Kunhwi; Xu, Hao

    2016-08-12

    Within the Natural Barrier System (NBS) group of the Used Fuel Disposition (UFD) Campaign at the Department of Energy’s (DOE) Office of Nuclear Energy, LBNL’s research activities have focused on understanding and modeling EDZ evolution and the associated coupled processes and impacts of high temperature on parameters and processes relevant to performance of a clay repository to establish the technical base for the maximum allowable temperature. This report documents results from some of these activities. These activities address key Features, Events, and Processes (FEPs), which have been ranked in importance from medium to high, as listed in Table 7 ofmore » the Used Fuel Disposition Campaign Disposal Research and Development Roadmap (FCR&D-USED-2011-000065 REV0) (Nutt, 2011). Specifically, they address FEP 2.2.01, Excavation Disturbed Zone, for clay/shale, by investigating how coupled processes affect EDZ evolution; FEP 2.2.05, Flow and Transport Pathways; and FEP 2.2.08, Hydrologic Processes, and FEP 2.2.07, Mechanical Processes, and FEP 2.2.09, Chemical Process—Transport, by studying near-field coupled THMC processes in clay/shale repositories. The activities documented in this report also address a number of research topics identified in Research & Development (R&D) Plan for Used Fuel Disposition Campaign (UFDC) Natural System Evaluation and Tool Development (Wang 2011), including Topics S3, Disposal system modeling – Natural System; P1, Development of discrete fracture network (DFN) model; P14, Technical basis for thermal loading limits; and P15 Modeling of disturbed rock zone (DRZ) evolution (clay repository).« less

  12. Where Will All Your Samples Go?

    NASA Astrophysics Data System (ADS)

    Lehnert, K.

    2017-12-01

    Even in the digital age, physical samples remain an essential component of Earth and space science research. Geoscientists collect samples, sometimes locally, often in remote locations during expensive field expeditions, or at sample repositories and museums. They take these samples to their labs to describe and analyze them. When the analyses are completed and the results are published, the samples get stored away in sheds, basements, or desk drawers, where they remain unknown and inaccessible to the broad science community. In some cases, they will get re-analyzed or shared with other researchers, who know of their existence through personal connections. The sad end comes when the researcher retires: There are many stories of samples and entire collections being discarded to free up space for new samples or other purposes, even though these samples may be unique and irreplaceable. Institutions do not feel obligated and do not have the resources to store samples in perpetuity. Only samples collected in large sampling campaigns such as the Ocean Discovery Program or cores taken on ships find a home in repositories that curate and preserve them for reuse in future science endeavors. In the era of open, transparent, and reproducible science, preservation and persistent access to samples must be considered a mandate. Policies need to be developed that guide investigators, institutions, and funding agencies to plan and implement solutions for reliably and persistently curating and providing access to samples. Registration of samples in online catalogs and use of persistent identifiers such as the International Geo Sample Number are first steps to ensure discovery and access of samples. But digital discovery and access loses its value if the physical objects are not preserved and accessible. It is unreasonable to expect that every sample ever collected can be archived. Selection of those samples that are worth preserving requires guidelines and policies. We also need to define standards that institutions must comply with to function as a trustworthy sample repository similar to trustworthy digital repositories. The iSamples Research Coordination Network of the EarthCube program aims to address some of these questions in workshops planned for 2018. This panel session offers an opportunity to ignite the discussion.

  13. The United States Antarctic Program Data Center (USAP-DC): Recent Developments

    NASA Astrophysics Data System (ADS)

    Nitsche, F. O.; Bauer, R.; Arko, R. A.; Shane, N.; Carbotte, S. M.; Scambos, T.

    2017-12-01

    Antarctic earth and environmental science data are highly valuable, often unique research assets. They are acquired with substantial and expensive logistical effort, frequently in areas that will not be re-visited for many years. The data acquired in support of Antarctic research span a wide range of disciplines. Historically, data management for the US Antarctic Program (USAP) has made use of existing disciplinary data centers, and the international Antarctic Master Directory (AMD) has served as a central metadata catalog linking to data files hosted in these external repositories. However, disciplinary repositories do not exist for all USAP-generated data types and often it is unclear what repositories are appropriate, leading to many datasets being served locally from scientist's websites or not available at all. The USAP Data Center (USAP-DC; www.usap-dc.org), operated as part of the Interdisciplinary Earth Data Alliance (IEDA), contributes to the broader preservation of research data acquired with funding from NSF's Office of Polar Programs by providing a repository for diverse data from the Antarctic region. USAP-DC hosts data that spans the range of Antarctic research from snow radar to volcano observatory imagery to penguin counts to meteorological model outputs. Data services include data documentation, long-term preservation, and web publication, as well as scientist support for registration of data descriptions into the AMD in fulfillment of US obligations under the International Antarctic Treaty. In Spring 2016, USAP-DC and the NSIDC began a new collaboration to consolidate data services for Antarctic investigators and to integrate the NSF-funded glaciology collection at NSIDC with the collection hosted by USAP-DC. Investigator submissions for NSF's Glaciology program now make use of USAP-DC's web submission tools, providing a uniform interface for Antarctic investigators. The tools have been redesigned to collect a broader range of metadata. Each data submission is reviewed and verified by a specialist from the USAP-DC/NSIDC team depending on disciplinary focus of the submission. A recently updated web search interface is available to search data by title, NSF program, award, dataset contributor, large scale project (e.g. WAIS Divide Ice Core) or by specifying an area in map view.

  14. A digital repository with an extensible data model for biobanking and genomic analysis management.

    PubMed

    Izzo, Massimiliano; Mortola, Francesco; Arnulfo, Gabriele; Fato, Marco M; Varesio, Luigi

    2014-01-01

    Molecular biology laboratories require extensive metadata to improve data collection and analysis. The heterogeneity of the collected metadata grows as research is evolving in to international multi-disciplinary collaborations and increasing data sharing among institutions. Single standardization is not feasible and it becomes crucial to develop digital repositories with flexible and extensible data models, as in the case of modern integrated biobanks management. We developed a novel data model in JSON format to describe heterogeneous data in a generic biomedical science scenario. The model is built on two hierarchical entities: processes and events, roughly corresponding to research studies and analysis steps within a single study. A number of sequential events can be grouped in a process building up a hierarchical structure to track patient and sample history. Each event can produce new data. Data is described by a set of user-defined metadata, and may have one or more associated files. We integrated the model in a web based digital repository with a data grid storage to manage large data sets located in geographically distinct areas. We built a graphical interface that allows authorized users to define new data types dynamically, according to their requirements. Operators compose queries on metadata fields using a flexible search interface and run them on the database and on the grid. We applied the digital repository to the integrated management of samples, patients and medical history in the BIT-Gaslini biobank. The platform currently manages 1800 samples of over 900 patients. Microarray data from 150 analyses are stored on the grid storage and replicated on two physical resources for preservation. The system is equipped with data integration capabilities with other biobanks for worldwide information sharing. Our data model enables users to continuously define flexible, ad hoc, and loosely structured metadata, for information sharing in specific research projects and purposes. This approach can improve sensitively interdisciplinary research collaboration and allows to track patients' clinical records, sample management information, and genomic data. The web interface allows the operators to easily manage, query, and annotate the files, without dealing with the technicalities of the data grid.

  15. A digital repository with an extensible data model for biobanking and genomic analysis management

    PubMed Central

    2014-01-01

    Motivation Molecular biology laboratories require extensive metadata to improve data collection and analysis. The heterogeneity of the collected metadata grows as research is evolving in to international multi-disciplinary collaborations and increasing data sharing among institutions. Single standardization is not feasible and it becomes crucial to develop digital repositories with flexible and extensible data models, as in the case of modern integrated biobanks management. Results We developed a novel data model in JSON format to describe heterogeneous data in a generic biomedical science scenario. The model is built on two hierarchical entities: processes and events, roughly corresponding to research studies and analysis steps within a single study. A number of sequential events can be grouped in a process building up a hierarchical structure to track patient and sample history. Each event can produce new data. Data is described by a set of user-defined metadata, and may have one or more associated files. We integrated the model in a web based digital repository with a data grid storage to manage large data sets located in geographically distinct areas. We built a graphical interface that allows authorized users to define new data types dynamically, according to their requirements. Operators compose queries on metadata fields using a flexible search interface and run them on the database and on the grid. We applied the digital repository to the integrated management of samples, patients and medical history in the BIT-Gaslini biobank. The platform currently manages 1800 samples of over 900 patients. Microarray data from 150 analyses are stored on the grid storage and replicated on two physical resources for preservation. The system is equipped with data integration capabilities with other biobanks for worldwide information sharing. Conclusions Our data model enables users to continuously define flexible, ad hoc, and loosely structured metadata, for information sharing in specific research projects and purposes. This approach can improve sensitively interdisciplinary research collaboration and allows to track patients' clinical records, sample management information, and genomic data. The web interface allows the operators to easily manage, query, and annotate the files, without dealing with the technicalities of the data grid. PMID:25077808

  16. Perspectives of the optical coherence tomography community on code and data sharing

    NASA Astrophysics Data System (ADS)

    Lurie, Kristen L.; Mistree, Behram F. T.; Ellerbee, Audrey K.

    2015-03-01

    As optical coherence tomography (OCT) grows to be a mature and successful field, it is important for the research community to develop a stronger practice of sharing code and data. A prolific culture of sharing can enable new and emerging laboratories to enter the field, allow research groups to gain new exposure and notoriety, and enable benchmarking of new algorithms and methods. Our long-term vision is to build tools to facilitate a stronger practice of sharing within this community. In line with this goal, our first aim was to understand the perceptions and practices of the community with respect to sharing research contributions (i.e., as code and data). We surveyed 52 members of the OCT community using an online polling system. Our main findings indicate that while researchers infrequently share their code and data, they are willing to contribute their research resources to a shared repository, and they believe that such a repository would benefit both their research and the OCT community at large. We plan to use the results of this survey to design a platform targeted to the OCT research community - an effort that ultimately aims to facilitate a more prolific culture of sharing.

  17. Credit where credit is due: indexing and exposing data citations in international data repository networks

    NASA Astrophysics Data System (ADS)

    Jones, M. B.; Vieglais, D.; Cruse, P.; Chodacki, J.; Budden, A. E.; Fenner, M.; Lowenberg, D.; Abrams, S.

    2017-12-01

    Research data are fundamental to the success of the academic enterprise, and yet the practice of citing data in academic and applied works is not widespread among researchers. Researchers need credit for their contributions, and yet current citation infrastructure focuses primarily on citations to research literature. Some citation indiexing systems even systematically exclude citations to data from their corpus. The Making Data Count (MDC) project will enable measuring the impact of research data much as is currently being done with publications, the primary vehicle for scholarly credit and accountability. The MDC team (including the California Digital Library, COUNTER, DataCite, and DataONE) are working together to publish a new COUNTER recommendation on data usage statistics; launch a DataCite-hosted MDC service for aggregated DLM based on the open-source Lagotto platform; and build tools for data repository and discovery services to easily integrate with the new MDC service. In providing such data-level metrics (DLM), the MDC project augments existing measures of scholarly success and so offers an important incentive promoting open data principles and quality research data through adoption of research data management best practices.

  18. 48 CFR 227.7207 - Contractor data repositories.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... Computer Software and Computer Software Documentation 227.7207 Contractor data repositories. Follow 227.7108 when it is in the Government's interests to have a data repository include computer software or to have a separate computer software repository. Contractual instruments establishing the repository...

  19. 48 CFR 227.7207 - Contractor data repositories.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... Computer Software and Computer Software Documentation 227.7207 Contractor data repositories. Follow 227.7108 when it is in the Government's interests to have a data repository include computer software or to have a separate computer software repository. Contractual instruments establishing the repository...

  20. 48 CFR 227.7207 - Contractor data repositories.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... Computer Software and Computer Software Documentation 227.7207 Contractor data repositories. Follow 227.7108 when it is in the Government's interests to have a data repository include computer software or to have a separate computer software repository. Contractual instruments establishing the repository...

  1. 48 CFR 227.7207 - Contractor data repositories.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... Computer Software and Computer Software Documentation 227.7207 Contractor data repositories. Follow 227.7108 when it is in the Government's interests to have a data repository include computer software or to have a separate computer software repository. Contractual instruments establishing the repository...

  2. Open Content in Open Context

    ERIC Educational Resources Information Center

    Kansa, Sarah Whitcher; Kansa, Eric C.

    2007-01-01

    This article presents the challenges and rewards of sharing research content through a discussion of Open Context, a new open access data publication system for field sciences and museum collections. Open Context is the first data repository of its kind, allowing self-publication of research data, community commentary through tagging, and clear…

  3. Understanding Teacher Professional Learning through Cyber Research

    ERIC Educational Resources Information Center

    Bates, Meg S.; Phalen, Lena; Moran, Cheryl

    2018-01-01

    Online professional learning websites provide a unique window into how teachers make self-directed choices about their own professional development. This study extends previous research on how teachers use online resource repositories to examine how teachers make choices about resource use on a professional learning website. The website, the…

  4. mSciences: An Affinity Space for Science Teachers

    ERIC Educational Resources Information Center

    Mota, Jorge; Morais, Carla; Moreira, Luciano; Paiva, João C.

    2017-01-01

    The project "Multimedia in science teaching: five years of research and teaching in Portugal" was successful in featuring the national research on multimedia in science education and in providing the community with a simple reference tool--a repository of open access scientific texts. The current work aims to describe the theoretical…

  5. Morphological and physio-chemical characterization of five Canistel accessions at the subtropical horticulture research station in Miami Florida

    USDA-ARS?s Scientific Manuscript database

    Fruit of five canistel cultivars, 'Fairchild','E11', 'Keisau', 'TREC#3' and 'TREC 3680' were evaluated and characterized at the National Germplasm Repository, Subtropical horticulture Research Station (SHRS) Miami, Florida. Thirty fruits were harvested from clonal accessions during July and August, ...

  6. Methods of Quality Appraisal for Studies Reviewed by Evidence Clearinghouses

    ERIC Educational Resources Information Center

    Wilson, Sandra Jo; Tanner-Smith, Emily

    2015-01-01

    This presentation will discuss quality appraisal methods for assessing research studies used in systematic reviews, research syntheses, and evidence-based practice repositories such as the What Works Clearinghouse. The different ways that the methodological rigor and risk of bias of primary studies included in syntheses is assessed means that…

  7. Accelerated Weathering of Fluidized Bed Steam Reformation Material Under Hydraulically Unsaturated Conditions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pierce, Eric M.

    2007-09-16

    To predict the long-term fate of low- and high-level waste forms in the subsurface over geologic time scales, it is important to understand the behavior of the corroding waste forms under conditions the mimic to the open flow and transport properties of a subsurface repository. Fluidized bed steam reformation (FBSR), a supplemental treatment technology option, is being considered as a waste form for the immobilization of low-activity tank waste. To obtain the fundamental information needed to evaluate the behavior of the FBSR waste form under repository relevant conditions and to monitor the long-term behavior of this material, an accelerated weatheringmore » experiment is being conducted with the pressurized unsaturated flow (PUF) apparatus. Unlike other accelerated weathering test methods (product consistency test, vapor hydration test, and drip test), PUF experiments are conducted under hydraulically unsaturated conditions. These experiments are unique because they mimic the vadose zone environment and allow the corroding waste form to achieve its final reaction state. Results from this on-going experiment suggest the volumetric water content varied as a function of time and reached steady state after 160 days of testing. Unlike the volumetric water content, periodic excursions in the solution pH and electrical conductivity have been occurring consistently during the test. Release of elements from the column illustrates a general trend of decreasing concentration with increasing reaction time. Normalized concentrations of K, Na, P, Re (a chemical analogue for 99Tc), and S are as much as 1 × 104 times greater than Al, Cr, Si, and Ti. After more than 600 days of testing, the solution chemistry data collected to-date illustrate the importance of understanding the long-term behavior of the FBSR product under conditions that mimic the open flow and transport properties of a subsurface repository.« less

  8. Health professional learner attitudes and use of digital learning resources.

    PubMed

    Maloney, Stephen; Chamberlain, Michael; Morrison, Shane; Kotsanas, George; Keating, Jennifer L; Ilic, Dragan

    2013-01-16

    Web-based digital repositories allow educational resources to be accessed efficiently and conveniently from diverse geographic locations, hold a variety of resource formats, enable interactive learning, and facilitate targeted access for the user. Unlike some other learning management systems (LMS), resources can be retrieved through search engines and meta-tagged labels, and content can be streamed, which is particularly useful for multimedia resources. The aim of this study was to examine usage and user experiences of an online learning repository (Physeek) in a population of physiotherapy students. The secondary aim of this project was to examine how students prefer to access resources and which resources they find most helpful. The following data were examined using an audit of the repository server: (1) number of online resources accessed per day in 2010, (2) number of each type of resource accessed, (3) number of resources accessed during business hours (9 am to 5 pm) and outside business hours (years 1-4), (4) session length of each log-on (years 1-4), and (5) video quality (bit rate) of each video accessed. An online questionnaire and 3 focus groups assessed student feedback and self-reported experiences of Physeek. Students preferred the support provided by Physeek to other sources of educational material primarily because of its efficiency. Peak usage commonly occurred at times of increased academic need (ie, examination times). Students perceived online repositories as a potential tool to support lifelong learning and health care delivery. The results of this study indicate that today's health professional students welcome the benefits of online learning resources because of their convenience and usability. This represents a transition away from traditional learning styles and toward technological learning support and may indicate a growing link between social immersions in Internet-based connections and learning styles. The true potential for Web-based resources to support student learning is as yet unknown.

  9. Health Professional Learner Attitudes and Use of Digital Learning Resources

    PubMed Central

    Chamberlain, Michael; Morrison, Shane; Kotsanas, George; Keating, Jennifer L; Ilic, Dragan

    2013-01-01

    Background Web-based digital repositories allow educational resources to be accessed efficiently and conveniently from diverse geographic locations, hold a variety of resource formats, enable interactive learning, and facilitate targeted access for the user. Unlike some other learning management systems (LMS), resources can be retrieved through search engines and meta-tagged labels, and content can be streamed, which is particularly useful for multimedia resources. Objective The aim of this study was to examine usage and user experiences of an online learning repository (Physeek) in a population of physiotherapy students. The secondary aim of this project was to examine how students prefer to access resources and which resources they find most helpful. Methods The following data were examined using an audit of the repository server: (1) number of online resources accessed per day in 2010, (2) number of each type of resource accessed, (3) number of resources accessed during business hours (9 am to 5 pm) and outside business hours (years 1-4), (4) session length of each log-on (years 1-4), and (5) video quality (bit rate) of each video accessed. An online questionnaire and 3 focus groups assessed student feedback and self-reported experiences of Physeek. Results Students preferred the support provided by Physeek to other sources of educational material primarily because of its efficiency. Peak usage commonly occurred at times of increased academic need (ie, examination times). Students perceived online repositories as a potential tool to support lifelong learning and health care delivery. Conclusions The results of this study indicate that today’s health professional students welcome the benefits of online learning resources because of their convenience and usability. This represents a transition away from traditional learning styles and toward technological learning support and may indicate a growing link between social immersions in Internet-based connections and learning styles. The true potential for Web-based resources to support student learning is as yet unknown. PMID:23324800

  10. Laboratory E-Notebooks: A Learning Object-Based Repository

    ERIC Educational Resources Information Center

    Abari, Ilior; Pierre, Samuel; Saliah-Hassane, Hamadou

    2006-01-01

    During distributed virtual laboratory experiment sessions, a major problem is to be able to collect, store, manage and share heterogeneous data (intermediate results, analysis, annotations, etc) manipulated simultaneously by geographically distributed teammates composing a virtual team. The electronic notebook is a possible response to this…

  11. Fast and Accurate Metadata Authoring Using Ontology-Based Recommendations.

    PubMed

    Martínez-Romero, Marcos; O'Connor, Martin J; Shankar, Ravi D; Panahiazar, Maryam; Willrett, Debra; Egyedi, Attila L; Gevaert, Olivier; Graybeal, John; Musen, Mark A

    2017-01-01

    In biomedicine, high-quality metadata are crucial for finding experimental datasets, for understanding how experiments were performed, and for reproducing those experiments. Despite the recent focus on metadata, the quality of metadata available in public repositories continues to be extremely poor. A key difficulty is that the typical metadata acquisition process is time-consuming and error prone, with weak or nonexistent support for linking metadata to ontologies. There is a pressing need for methods and tools to speed up the metadata acquisition process and to increase the quality of metadata that are entered. In this paper, we describe a methodology and set of associated tools that we developed to address this challenge. A core component of this approach is a value recommendation framework that uses analysis of previously entered metadata and ontology-based metadata specifications to help users rapidly and accurately enter their metadata. We performed an initial evaluation of this approach using metadata from a public metadata repository.

  12. Fast and Accurate Metadata Authoring Using Ontology-Based Recommendations

    PubMed Central

    Martínez-Romero, Marcos; O’Connor, Martin J.; Shankar, Ravi D.; Panahiazar, Maryam; Willrett, Debra; Egyedi, Attila L.; Gevaert, Olivier; Graybeal, John; Musen, Mark A.

    2017-01-01

    In biomedicine, high-quality metadata are crucial for finding experimental datasets, for understanding how experiments were performed, and for reproducing those experiments. Despite the recent focus on metadata, the quality of metadata available in public repositories continues to be extremely poor. A key difficulty is that the typical metadata acquisition process is time-consuming and error prone, with weak or nonexistent support for linking metadata to ontologies. There is a pressing need for methods and tools to speed up the metadata acquisition process and to increase the quality of metadata that are entered. In this paper, we describe a methodology and set of associated tools that we developed to address this challenge. A core component of this approach is a value recommendation framework that uses analysis of previously entered metadata and ontology-based metadata specifications to help users rapidly and accurately enter their metadata. We performed an initial evaluation of this approach using metadata from a public metadata repository. PMID:29854196

  13. Experiments and Modeling to Support Field Test Design

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Johnson, Peter Jacob; Bourret, Suzanne Michelle; Zyvoloski, George Anthony

    Disposition of heat-generating nuclear waste (HGNW) remains a continuing technical and sociopolitical challenge. We define HGNW as the combination of both heat generating defense high level waste (DHLW) and civilian spent nuclear fuel (SNF). Numerous concepts for HGNW management have been proposed and examined internationally, including an extensive focus on geologic disposal (c.f. Brunnengräber et al., 2013). One type of proposed geologic material is salt, so chosen because of its viscoplastic deformation that causes self-repair of damage or deformation induced in the salt by waste emplacement activities (Hansen and Leigh, 2011). Salt as a repository material has been tested atmore » several sites around the world, notably the Morsleben facility in Germany (c.f. Fahland and Heusermann, 2013; Wollrath et al., 2014; Fahland et al., 2015) and at the Waste Isolation Pilot Plant (WIPP) near Carlsbad, NM. Evaluating the technical feasibility of a HGNW repository in salt is an ongoing process involving experiments and numerical modeling of many processes at many facilities.« less

  14. 40 CFR 124.33 - Information repository.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 21 2010-07-01 2010-07-01 false Information repository. 124.33 Section... FOR DECISIONMAKING Specific Procedures Applicable to RCRA Permits § 124.33 Information repository. (a... basis, for an information repository. When assessing the need for an information repository, the...

  15. 10 CFR 60.130 - General considerations.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... REPOSITORIES Technical Criteria Design Criteria for the Geologic Repository Operations Area § 60.130 General... for a high-level radioactive waste repository at a geologic repository operations area, and an... geologic repository operations area, must include the principal design criteria for a proposed facility...

  16. Designing Research Services: Cross-Disciplinary Administration and the Research Lifecycle

    NASA Astrophysics Data System (ADS)

    Madden, G.

    2017-12-01

    The sheer number of technical and administrative offices involved in the research lifecycle, and the lack of shared governance and shared processes across those offices, creates challenges to the successful preservation of research outputs. Universities need a more integrated approach to the research lifecycle that allows us to: recognize a research project as it is being initiated; identify the data associated with the research project; document and track any compliance, security, access, and publication requirements associated with the research and its data; follow the research and its associated components across the research lifecycle; and finally recognize that the research has come to a close so we can trigger the various preservation, access, and communications processes that close the loop, inform the public, and promote the continued progress of science. Such an approach will require cooperation, communications, and shared workflow tools that tie together (often across many years) PIs, research design methodologists, grants offices, contract negotiators, central research administrators, research compliance specialists, desktop IT support units, server administrators, high performance computing facilities, data centers, specialized data transfer networks, institutional research repositories, institutional data repositories, and research communications groups, all of which play a significant role in the technical or administrative success of research. This session will focus on progress towards improving cross-disciplinary administrative and technical cooperation at Penn State University, with an emphasis on generalizable approaches that can be adopted elsewhere.

  17. A laboratory validation study of the time-lapse oscillatory pumping test concept for leakage detection in geological repositories

    NASA Astrophysics Data System (ADS)

    Sun, A. Y.; Islam, A.; Lu, J.

    2017-12-01

    Time-lapse oscillatory pumping test (OPT) has been introduced recently as a pressure-based monitoring technique for detecting potential leakage in geologic repositories. By routinely conducting OPT at a number of pulsing frequencies, a site operator may identify the potential anomalies in the frequency domain, alleviating the ambiguity caused by reservoir noise and improving the signal-to-noise ratio. Building on previous theoretical and field studies, this work performed a series of laboratory experiments to validate the concept of time-lapse OPT using a custom made, stainless steel tank under relatively high pressures ( 120psi). The experimental configuration simulates a miniature geologic storage repository consisting of three layers (i.e., injection zone, caprock, and above-zone aquifer). Results show that leakage in the injection zone led to deviations in the power spectrum of observed pressure data, and the amplitude of which also increases with decreasing pulsing frequencies. The experimental results were further analyzed by developing a 3D flow model, using which the model parameters were estimated through frequency domain inversion.

  18. 48 CFR 227.7108 - Contractor data repositories.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... Technical Data 227.7108 Contractor data repositories. (a) Contractor data repositories may be established... procedures for protecting technical data delivered to or stored at the repository from unauthorized release... disclosure of technical data from the repository to third parties consistent with the Government's rights in...

  19. DSA-WDS Common Requirements: Developing a New Core Data Repository Certification

    NASA Astrophysics Data System (ADS)

    Minster, J. B. H.; Edmunds, R.; L'Hours, H.; Mokrane, M.; Rickards, L.

    2016-12-01

    The Data Seal of Approval (DSA) and the International Council for Science - World Data System (ICSU-WDS) have both developed minimally intensive core certification standards whereby digital repositories supply evidence that they are trustworthy and have a long-term outlook. Both DSA and WDS applicants have found core certification to be beneficial: building stakeholder confidence, enhancing the repository's reputation, and demonstrating that it is following good practices; as well as stimulating the repository to focus on processes and procedures, thereby achieving ever higher levels of professionalism over time.The DSA and WDS core certifications evolved independently serving initially different communities but both initiatives are multidisciplinary with catalogues of criteria and review procedures based on the same principles. Hence, to realize efficiencies, simplify assessment options, stimulate more certifications, and increase impact on the community, the Repository Audit and Certification DSA-WDS Partnership Working Group (WG) was established under the umbrella of the Research Data Alliance (RDA). The WG conducted a side-by-side analysis of both frameworks to unify the wording and criteria, ultimately leading to a harmonized Catalogue of Common Requirements for core certification of repositories—as well as a set of Common Procedures for their assessment.This presentation will focus on the collaborative effort by DSA and WDS to establish (1) a testbed comprising DSA and WDS certified data repositories to validate both the new Catalogue and Procedures, and (2) a joint Certification Board towards their practical implementation. We will describe:• The purpose and methodology of the testbed, including selection of repositories to be assessed against the common standard.• The results of the testbed, with an in-depth look at some of the comments received and issues highlighted.• General insights gained from evaluating the testbed results, the subsequent changes to the Common Requirements and Procedures, and an assessment of the success of these enhancements.• Steps by the two organizations to integrate the Common Certification into their tools and systems. In particular, the creation of Terms of Reference for the nascent DSA-WDS Certification Board.

  20. Design and application of a data-independent precursor and product ion repository.

    PubMed

    Thalassinos, Konstantinos; Vissers, Johannes P C; Tenzer, Stefan; Levin, Yishai; Thompson, J Will; Daniel, David; Mann, Darrin; DeLong, Mark R; Moseley, M Arthur; America, Antoine H; Ottens, Andrew K; Cavey, Greg S; Efstathiou, Georgios; Scrivens, James H; Langridge, James I; Geromanos, Scott J

    2012-10-01

    The functional design and application of a data-independent LC-MS precursor and product ion repository for protein identification, quantification, and validation is conceptually described. The ion repository was constructed from the sequence search results of a broad range of discovery experiments investigating various tissue types of two closely related mammalian species. The relative high degree of similarity in protein complement, ion detection, and peptide and protein identification allows for the analysis of normalized precursor and product ion intensity values, as well as standardized retention times, creating a multidimensional/orthogonal queryable, qualitative, and quantitative space. Peptide ion map selection for identification and quantification is primarily based on replication and limited variation. The information is stored in a relational database and is used to create peptide- and protein-specific fragment ion maps that can be queried in a targeted fashion against the raw or time aligned ion detections. These queries can be conducted either individually or as groups, where the latter affords pathway and molecular machinery analysis of the protein complement. The presented results also suggest that peptide ionization and fragmentation efficiencies are highly conserved between experiments and practically independent of the analyzed biological sample when using similar instrumentation. Moreover, the data illustrate only minor variation in ionization efficiency with amino acid sequence substitutions occurring between species. Finally, the data and the presented results illustrate how LC-MS performance metrics can be extracted and utilized to ensure optimal performance of the employed analytical workflows.

Top