Sample records for services api tools

  1. API REST Web service and backend system Of Lecturer’s Assessment Information System on Politeknik Negeri Bali

    NASA Astrophysics Data System (ADS)

    Manuaba, I. B. P.; Rudiastini, E.

    2018-01-01

    Assessment of lecturers is a tool used to measure lecturer performance. Lecturer’s assessment variable can be measured from three aspects : teaching activities, research and community service. Broad aspect to measure the performance of lecturers requires a special framework, so that the system can be developed in a sustainable manner. Issues of this research is to create a API web service data tool, so the lecturer assessment system can be developed in various frameworks. The research was developed with web service and php programming language with the output of json extension data. The conclusion of this research is API web service data application can be developed using several platforms such as web, mobile application

  2. The APIS service : a tool for accessing value-added HST planetary auroral observations over 1997-2015

    NASA Astrophysics Data System (ADS)

    Lamy, L.; Henry, F.; Prangé, R.; Le Sidaner, P.

    2015-10-01

    The Auroral Planetary Imaging and Spectroscopy (APIS) service http://obspm.fr/apis/ provides an open and interactive access to processed auroral observations of the outer planets and their satellites. Such observations are of interest for a wide community at the interface between planetology, magnetospheric and heliospheric physics. APIS consists of (i) a high level database, built from planetary auroral observations acquired by the Hubble Space Telescope (HST) since 1997 with its mostly used Far-Ultraviolet spectro- imagers, (ii) a dedicated search interface aimed at browsing efficiently this database through relevant conditional search criteria (Figure 1) and (iii) the ability to interactively work with the data online through plotting tools developed by the Virtual Observatory (VO) community, such as Aladin and Specview. This service is VO compliant and can therefore also been queried by external search tools of the VO community. The diversity of available data and the capability to sort them out by relevant physical criteria shall in particular facilitate statistical studies, on long-term scales and/or multi-instrumental multispectral combined analysis [1,2]. We will present the updated capabilities of APIS with several examples. Several tutorials are available online.

  3. A RESTful API for accessing microbial community data for MG-RAST.

    PubMed

    Wilke, Andreas; Bischof, Jared; Harrison, Travis; Brettin, Tom; D'Souza, Mark; Gerlach, Wolfgang; Matthews, Hunter; Paczian, Tobias; Wilkening, Jared; Glass, Elizabeth M; Desai, Narayan; Meyer, Folker

    2015-01-01

    Metagenomic sequencing has produced significant amounts of data in recent years. For example, as of summer 2013, MG-RAST has been used to annotate over 110,000 data sets totaling over 43 Terabases. With metagenomic sequencing finding even wider adoption in the scientific community, the existing web-based analysis tools and infrastructure in MG-RAST provide limited capability for data retrieval and analysis, such as comparative analysis between multiple data sets. Moreover, although the system provides many analysis tools, it is not comprehensive. By opening MG-RAST up via a web services API (application programmers interface) we have greatly expanded access to MG-RAST data, as well as provided a mechanism for the use of third-party analysis tools with MG-RAST data. This RESTful API makes all data and data objects created by the MG-RAST pipeline accessible as JSON objects. As part of the DOE Systems Biology Knowledgebase project (KBase, http://kbase.us) we have implemented a web services API for MG-RAST. This API complements the existing MG-RAST web interface and constitutes the basis of KBase's microbial community capabilities. In addition, the API exposes a comprehensive collection of data to programmers. This API, which uses a RESTful (Representational State Transfer) implementation, is compatible with most programming environments and should be easy to use for end users and third parties. It provides comprehensive access to sequence data, quality control results, annotations, and many other data types. Where feasible, we have used standards to expose data and metadata. Code examples are provided in a number of languages both to show the versatility of the API and to provide a starting point for users. We present an API that exposes the data in MG-RAST for consumption by our users, greatly enhancing the utility of the MG-RAST service.

  4. A RESTful API for accessing microbial community data for MG-RAST

    DOE PAGES

    Wilke, Andreas; Bischof, Jared; Harrison, Travis; ...

    2015-01-08

    Metagenomic sequencing has produced significant amounts of data in recent years. For example, as of summer 2013, MGRAST has been used to annotate over 110,000 data sets totaling over 43 Terabases. With metagenomic sequencing finding even wider adoption in the scientific community, the existing web-based analysis tools and infrastructure in MG-RAST provide limited capability for data retrieval and analysis, such as comparative analysis between multiple data sets. Moreover, although the system provides many analysis tools, it is not comprehensive. By opening MG-RAST up via a web services API (application programmers interface) we have greatly expanded access to MG-RAST data, asmore » well as provided a mechanism for the use of third-party analysis tools with MG-RAST data. This RESTful API makes all data and data objects created by the MG-RAST pipeline accessible as JSON objects. As part of the DOE Systems Biology Knowledgebase project (KBase, http:// kbase.us) we have implemented a web services API for MG-RAST. This API complements the existing MG-RAST web interface and constitutes the basis of KBase’s microbial community capabilities. In addition, the API exposes a comprehensive collection of data to programmers. This API, which uses a RESTful (Representational State Transfer) implementation, is compatible with most programming environments and should be easy to use for end users and third parties. It provides comprehensive access to sequence data, quality control results, annotations, and many other data types. Where feasible, we have used standards to expose data and metadata. Code examples are provided in a number of languages both to show the versatility of the API and to provide a starting point for users. We present an API that exposes the data in MG-RAST for consumption by our users, greatly enhancing the utility of the MG-RAST service.« less

  5. A RESTful API for Accessing Microbial Community Data for MG-RAST

    PubMed Central

    Wilke, Andreas; Bischof, Jared; Harrison, Travis; Brettin, Tom; D'Souza, Mark; Gerlach, Wolfgang; Matthews, Hunter; Paczian, Tobias; Wilkening, Jared; Glass, Elizabeth M.; Desai, Narayan; Meyer, Folker

    2015-01-01

    Metagenomic sequencing has produced significant amounts of data in recent years. For example, as of summer 2013, MG-RAST has been used to annotate over 110,000 data sets totaling over 43 Terabases. With metagenomic sequencing finding even wider adoption in the scientific community, the existing web-based analysis tools and infrastructure in MG-RAST provide limited capability for data retrieval and analysis, such as comparative analysis between multiple data sets. Moreover, although the system provides many analysis tools, it is not comprehensive. By opening MG-RAST up via a web services API (application programmers interface) we have greatly expanded access to MG-RAST data, as well as provided a mechanism for the use of third-party analysis tools with MG-RAST data. This RESTful API makes all data and data objects created by the MG-RAST pipeline accessible as JSON objects. As part of the DOE Systems Biology Knowledgebase project (KBase, http://kbase.us) we have implemented a web services API for MG-RAST. This API complements the existing MG-RAST web interface and constitutes the basis of KBase's microbial community capabilities. In addition, the API exposes a comprehensive collection of data to programmers. This API, which uses a RESTful (Representational State Transfer) implementation, is compatible with most programming environments and should be easy to use for end users and third parties. It provides comprehensive access to sequence data, quality control results, annotations, and many other data types. Where feasible, we have used standards to expose data and metadata. Code examples are provided in a number of languages both to show the versatility of the API and to provide a starting point for users. We present an API that exposes the data in MG-RAST for consumption by our users, greatly enhancing the utility of the MG-RAST service. PMID:25569221

  6. Accessing the SEED genome databases via Web services API: tools for programmers.

    PubMed

    Disz, Terry; Akhter, Sajia; Cuevas, Daniel; Olson, Robert; Overbeek, Ross; Vonstein, Veronika; Stevens, Rick; Edwards, Robert A

    2010-06-14

    The SEED integrates many publicly available genome sequences into a single resource. The database contains accurate and up-to-date annotations based on the subsystems concept that leverages clustering between genomes and other clues to accurately and efficiently annotate microbial genomes. The backend is used as the foundation for many genome annotation tools, such as the Rapid Annotation using Subsystems Technology (RAST) server for whole genome annotation, the metagenomics RAST server for random community genome annotations, and the annotation clearinghouse for exchanging annotations from different resources. In addition to a web user interface, the SEED also provides Web services based API for programmatic access to the data in the SEED, allowing the development of third-party tools and mash-ups. The currently exposed Web services encompass over forty different methods for accessing data related to microbial genome annotations. The Web services provide comprehensive access to the database back end, allowing any programmer access to the most consistent and accurate genome annotations available. The Web services are deployed using a platform independent service-oriented approach that allows the user to choose the most suitable programming platform for their application. Example code demonstrate that Web services can be used to access the SEED using common bioinformatics programming languages such as Perl, Python, and Java. We present a novel approach to access the SEED database. Using Web services, a robust API for access to genomics data is provided, without requiring large volume downloads all at once. The API ensures timely access to the most current datasets available, including the new genomes as soon as they come online.

  7. GIANT API: an application programming interface for functional genomics

    PubMed Central

    Roberts, Andrew M.; Wong, Aaron K.; Fisk, Ian; Troyanskaya, Olga G.

    2016-01-01

    GIANT API provides biomedical researchers programmatic access to tissue-specific and global networks in humans and model organisms, and associated tools, which includes functional re-prioritization of existing genome-wide association study (GWAS) data. Using tissue-specific interaction networks, researchers are able to predict relationships between genes specific to a tissue or cell lineage, identify the changing roles of genes across tissues and uncover disease-gene associations. Additionally, GIANT API enables computational tools like NetWAS, which leverages tissue-specific networks for re-prioritization of GWAS results. The web services covered by the API include 144 tissue-specific functional gene networks in human, global functional networks for human and six common model organisms and the NetWAS method. GIANT API conforms to the REST architecture, which makes it stateless, cacheable and highly scalable. It can be used by a diverse range of clients including web browsers, command terminals, programming languages and standalone apps for data analysis and visualization. The API is freely available for use at http://giant-api.princeton.edu. PMID:27098035

  8. Using Clouds for MapReduce Measurement Assignments

    ERIC Educational Resources Information Center

    Rabkin, Ariel; Reiss, Charles; Katz, Randy; Patterson, David

    2013-01-01

    We describe our experiences teaching MapReduce in a large undergraduate lecture course using public cloud services and the standard Hadoop API. Using the standard API, students directly experienced the quality of industrial big-data tools. Using the cloud, every student could carry out scalability benchmarking assignments on realistic hardware,…

  9. The Auroral Planetary Imaging and Spectroscopy (APIS) service

    NASA Astrophysics Data System (ADS)

    Lamy, L.; Prangé, R.; Henry, F.; Le Sidaner, P.

    2015-06-01

    The Auroral Planetary Imaging and Spectroscopy (APIS) service, accessible online, provides an open and interactive access to processed auroral observations of the outer planets and their satellites. Such observations are of interest for a wide community at the interface between planetology, magnetospheric and heliospheric physics. APIS consists of (i) a high level database, built from planetary auroral observations acquired by the Hubble Space Telescope (HST) since 1997 with its mostly used Far-Ultraviolet spectro-imagers, (ii) a dedicated search interface aimed at browsing efficiently this database through relevant conditional search criteria and (iii) the ability to interactively work with the data online through plotting tools developed by the Virtual Observatory (VO) community, such as Aladin and Specview. This service is VO compliant and can therefore also been queried by external search tools of the VO community. The diversity of available data and the capability to sort them out by relevant physical criteria shall in particular facilitate statistical studies, on long-term scales and/or multi-instrumental multi-spectral combined analysis.

  10. GIANT API: an application programming interface for functional genomics.

    PubMed

    Roberts, Andrew M; Wong, Aaron K; Fisk, Ian; Troyanskaya, Olga G

    2016-07-08

    GIANT API provides biomedical researchers programmatic access to tissue-specific and global networks in humans and model organisms, and associated tools, which includes functional re-prioritization of existing genome-wide association study (GWAS) data. Using tissue-specific interaction networks, researchers are able to predict relationships between genes specific to a tissue or cell lineage, identify the changing roles of genes across tissues and uncover disease-gene associations. Additionally, GIANT API enables computational tools like NetWAS, which leverages tissue-specific networks for re-prioritization of GWAS results. The web services covered by the API include 144 tissue-specific functional gene networks in human, global functional networks for human and six common model organisms and the NetWAS method. GIANT API conforms to the REST architecture, which makes it stateless, cacheable and highly scalable. It can be used by a diverse range of clients including web browsers, command terminals, programming languages and standalone apps for data analysis and visualization. The API is freely available for use at http://giant-api.princeton.edu. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  11. Using ChEMBL web services for building applications and data processing workflows relevant to drug discovery.

    PubMed

    Nowotka, Michał M; Gaulton, Anna; Mendez, David; Bento, A Patricia; Hersey, Anne; Leach, Andrew

    2017-08-01

    ChEMBL is a manually curated database of bioactivity data on small drug-like molecules, used by drug discovery scientists. Among many access methods, a REST API provides programmatic access, allowing the remote retrieval of ChEMBL data and its integration into other applications. This approach allows scientists to move from a world where they go to the ChEMBL web site to search for relevant data, to one where ChEMBL data can be simply integrated into their everyday tools and work environment. Areas covered: This review highlights some of the audiences who may benefit from using the ChEMBL API, and the goals they can address, through the description of several use cases. The examples cover a team communication tool (Slack), a data analytics platform (KNIME), batch job management software (Luigi) and Rich Internet Applications. Expert opinion: The advent of web technologies, cloud computing and micro services oriented architectures have made REST APIs an essential ingredient of modern software development models. The widespread availability of tools consuming RESTful resources have made them useful for many groups of users. The ChEMBL API is a valuable resource of drug discovery bioactivity data for professional chemists, chemistry students, data scientists, scientific and web developers.

  12. Enabling Mobile Air Quality App Development with an AirNow API

    NASA Astrophysics Data System (ADS)

    Dye, T.; White, J. E.; Ludewig, S. A.; Dickerson, P.; Healy, A. N.; West, J. W.; Prince, L. A.

    2013-12-01

    The U.S. Environmental Protection Agency's (EPA) AirNow program works with over 130 participating state, local, and federal air quality agencies to obtain, quality control, and store real-time air quality observations and forecasts. From these data, the AirNow system generates thousands of maps and products each hour. Each day, information from AirNow is published online and in other media to assist the public in making health-based decisions related to air quality. However, an increasing number of people use mobile devices as their primary tool for obtaining information, and AirNow has responded to this trend by publishing an easy-to-use Web API that is useful for mobile app developers. This presentation will describe the various features of the AirNow application programming interface (API), including Representational State Transfer (REST)-type web services, file outputs, and RSS feeds. In addition, a web portal for the AirNow API will be shown, including documentation on use of the system, a query tool for configuring and running web services, and general information about the air quality data and forecasts available. Data published via the AirNow API includes corresponding Air Quality Index (AQI) levels for each pollutant. We will highlight examples of mobile apps that are using the AirNow API to provide location-based, real-time air quality information. Examples will include mobile apps developed for Minnesota ('Minnesota Air') and Washington, D.C. ('Clean Air Partners Air Quality'), and an app developed by EPA ('EPA AirNow').

  13. Coordinating complex decision support activities across distributed applications

    NASA Technical Reports Server (NTRS)

    Adler, Richard M.

    1994-01-01

    Knowledge-based technologies have been applied successfully to automate planning and scheduling in many problem domains. Automation of decision support can be increased further by integrating task-specific applications with supporting database systems, and by coordinating interactions between such tools to facilitate collaborative activities. Unfortunately, the technical obstacles that must be overcome to achieve this vision of transparent, cooperative problem-solving are daunting. Intelligent decision support tools are typically developed for standalone use, rely on incompatible, task-specific representational models and application programming interfaces (API's), and run on heterogeneous computing platforms. Getting such applications to interact freely calls for platform independent capabilities for distributed communication, as well as tools for mapping information across disparate representations. Symbiotics is developing a layered set of software tools (called NetWorks! for integrating and coordinating heterogeneous distributed applications. he top layer of tools consists of an extensible set of generic, programmable coordination services. Developers access these services via high-level API's to implement the desired interactions between distributed applications.

  14. Scientific Platform as a Service - Tools and solutions for efficient access to and analysis of oceanographic data

    NASA Astrophysics Data System (ADS)

    Vines, Aleksander; Hansen, Morten W.; Korosov, Anton

    2017-04-01

    Existing infrastructure international and Norwegian projects, e.g., NorDataNet, NMDC and NORMAP, provide open data access through the OPeNDAP protocol following the conventions for CF (Climate and Forecast) metadata, designed to promote the processing and sharing of files created with the NetCDF application programming interface (API). This approach is now also being implemented in the Norwegian Sentinel Data Hub (satellittdata.no) to provide satellite EO data to the user community. Simultaneously with providing simplified and unified data access, these projects also seek to use and establish common standards for use and discovery metadata. This then allows development of standardized tools for data search and (subset) streaming over the internet to perform actual scientific analysis. A combinnation of software tools, which we call a Scientific Platform as a Service (SPaaS), will take advantage of these opportunities to harmonize and streamline the search, retrieval and analysis of integrated satellite and auxiliary observations of the oceans in a seamless system. The SPaaS is a cloud solution for integration of analysis tools with scientific datasets via an API. The core part of the SPaaS is a distributed metadata catalog to store granular metadata describing the structure, location and content of available satellite, model, and in situ datasets. The analysis tools include software for visualization (also online), interactive in-depth analysis, and server-based processing chains. The API conveys search requests between system nodes (i.e., interactive and server tools) and provides easy access to the metadata catalog, data repositories, and the tools. The SPaaS components are integrated in virtual machines, of which provisioning and deployment are automatized using existing state-of-the-art open-source tools (e.g., Vagrant, Ansible, Docker). The open-source code for scientific tools and virtual machine configurations is under version control at https://github.com/nansencenter/, and is coupled to an online continuous integration system (e.g., Travis CI).

  15. Earthdata Developer Portal poster

    NASA Astrophysics Data System (ADS)

    Plofchan, P.

    2016-12-01

    A common theme at community conferences in the Earth science domain is the need for more integration with related services, clearer documentation for services available, and a general simplification of what it takes to leverage existing tools so that setup and administration time can be minimized and time spent researching can be maximized. NASA's Earthdata Developer Portal (EDP) is the newly-created central location for documentation related to interacting with services offered by the EOSDIS community. The EDP will provide technical documentation for APIs, process documentation that provides real-world examples of how to use existing APIs in the real world, release notes for applications and services so that the entire community can stay up to date on recent updates, and best practice suggestions to improve implementation of both front-end and back-end services. Application and service owners will own their documentation while the EDP will ingest the documentation, serve it up in an interface using both industry standard tools, such as Swagger, and custom "adapters". The content is then styled to ensure consistency with other documentation found throughout the site and will be made searchable from a single location. "Getting Started" paths will also provide those users new to the space a simple path to follow to perform common tasks such as "searching and getting data" or "hosting an application on the Earthdata platform."

  16. SenseMyHeart: A cloud service and API for wearable heart monitors.

    PubMed

    Pinto Silva, P M; Silva Cunha, J P

    2015-01-01

    In the era of ubiquitous computing, the growing adoption of wearable systems and body sensor networks is trailing the path for new research and software for cardiovascular intensity, energy expenditure and stress and fatigue detection through cardiovascular monitoring. Several systems have received clinical-certification and provide huge amounts of reliable heart-related data in a continuous basis. PhysioNet provides equally reliable open-source software tools for ECG processing and analysis that can be combined with these devices. However, this software remains difficult to use in a mobile environment and for researchers unfamiliar with Linux-based systems. In the present paper we present an approach that aims at tackling these limitations by developing a cloud service that provides an API for a PhysioNet-based pipeline for ECG processing and Heart Rate Variability measurement. We describe the proposed solution, along with its advantages and tradeoffs. We also present some client tools (windows and Android) and several projects where the developed cloud service has been used successfully as a standard for Heart Rate and Heart Rate Variability studies in different scenarios.

  17. A Java API for working with PubChem datasets.

    PubMed

    Southern, Mark R; Griffin, Patrick R

    2011-03-01

    PubChem is a public repository of chemical structures and associated biological activities. The PubChem BioAssay database contains assay descriptions, conditions and readouts and biological screening results that have been submitted by the biomedical research community. The PubChem web site and Power User Gateway (PUG) web service allow users to interact with the data and raw files are available via FTP. These resources are helpful to many but there can also be great benefit by using a software API to manipulate the data. Here, we describe a Java API with entity objects mapped to the PubChem Schema and with wrapper functions for calling the NCBI eUtilities and PubChem PUG web services. PubChem BioAssays and associated chemical compounds can then be queried and manipulated in a local relational database. Features include chemical structure searching and generation and display of curve fits from stored dose-response experiments, something that is not yet available within PubChem itself. The aim is to provide researchers with a fast, consistent, queryable local resource from which to manipulate PubChem BioAssays in a database agnostic manner. It is not intended as an end user tool but to provide a platform for further automation and tools development. http://code.google.com/p/pubchemdb.

  18. Web Services--A Buzz Word with Potentials

    Treesearch

    János T. Füstös

    2006-01-01

    The simplest definition of a web service is an application that provides a web API. The web API exposes the functionality of the solution to other applications. The web API relies on other Internet-based technologies to manage communications. The resulting web services are pervasive, vendor-independent, language-neutral, and very low-cost. The main purpose of a web API...

  19. Developing Cancer Informatics Applications and Tools Using the NCI Genomic Data Commons API.

    PubMed

    Wilson, Shane; Fitzsimons, Michael; Ferguson, Martin; Heath, Allison; Jensen, Mark; Miller, Josh; Murphy, Mark W; Porter, James; Sahni, Himanso; Staudt, Louis; Tang, Yajing; Wang, Zhining; Yu, Christine; Zhang, Junjun; Ferretti, Vincent; Grossman, Robert L

    2017-11-01

    The NCI Genomic Data Commons (GDC) was launched in 2016 and makes available over 4 petabytes (PB) of cancer genomic and associated clinical data to the research community. This dataset continues to grow and currently includes over 14,500 patients. The GDC is an example of a biomedical data commons, which collocates biomedical data with storage and computing infrastructure and commonly used web services, software applications, and tools to create a secure, interoperable, and extensible resource for researchers. The GDC is (i) a data repository for downloading data that have been submitted to it, and also a system that (ii) applies a common set of bioinformatics pipelines to submitted data; (iii) reanalyzes existing data when new pipelines are developed; and (iv) allows users to build their own applications and systems that interoperate with the GDC using the GDC Application Programming Interface (API). We describe the GDC API and how it has been used both by the GDC itself and by third parties. Cancer Res; 77(21); e15-18. ©2017 AACR . ©2017 American Association for Cancer Research.

  20. A Java API for working with PubChem datasets

    PubMed Central

    Southern, Mark R.; Griffin, Patrick R.

    2011-01-01

    Summary: PubChem is a public repository of chemical structures and associated biological activities. The PubChem BioAssay database contains assay descriptions, conditions and readouts and biological screening results that have been submitted by the biomedical research community. The PubChem web site and Power User Gateway (PUG) web service allow users to interact with the data and raw files are available via FTP. These resources are helpful to many but there can also be great benefit by using a software API to manipulate the data. Here, we describe a Java API with entity objects mapped to the PubChem Schema and with wrapper functions for calling the NCBI eUtilities and PubChem PUG web services. PubChem BioAssays and associated chemical compounds can then be queried and manipulated in a local relational database. Features include chemical structure searching and generation and display of curve fits from stored dose–response experiments, something that is not yet available within PubChem itself. The aim is to provide researchers with a fast, consistent, queryable local resource from which to manipulate PubChem BioAssays in a database agnostic manner. It is not intended as an end user tool but to provide a platform for further automation and tools development. Availability: http://code.google.com/p/pubchemdb Contact: southern@scripps.edu PMID:21216779

  1. Earth Observation oriented teaching materials development based on OGC Web services and Bashyt generated reports

    NASA Astrophysics Data System (ADS)

    Stefanut, T.; Gorgan, D.; Giuliani, G.; Cau, P.

    2012-04-01

    Creating e-Learning materials in the Earth Observation domain is a difficult task especially for non-technical specialists who have to deal with distributed repositories, large amounts of information and intensive processing requirements. Furthermore, due to the lack of specialized applications for developing teaching resources, technical knowledge is required also for defining data presentation structures or in the development and customization of user interaction techniques for better teaching results. As a response to these issues during the GiSHEO FP7 project [1] and later in the EnviroGRIDS FP7 [2] project, we have developed the eGLE e-Learning Platform [3], a tool based application that provides dedicated functionalities to the Earth Observation specialists for developing teaching materials. The proposed architecture is built around a client-server design that provides the core functionalities (e.g. user management, tools integration, teaching materials settings, etc.) and has been extended with a distributed component implemented through the tools that are integrated into the platform, as described further. Our approach in dealing with multiple transfer protocol types, heterogeneous data formats or various user interaction techniques involve the development and integration of very specialized elements (tools) that can be customized by the trainers in a visual manner through simple user interfaces. In our concept each tool is dedicated to a specific data type, implementing optimized mechanisms for searching, retrieving, visualizing and interacting with it. At the same time, in each learning resource can be integrated any number of tools, through drag-and-drop interaction, allowing the teacher to retrieve pieces of data of various types (e.g. images, charts, tables, text, videos etc.) from different sources (e.g. OGC web services, charts created through Bashyt application, etc.) through different protocols (ex. WMS, BASHYT API, FTP, HTTP etc.) and to display them all together in a unitary manner using the same visual structure [4]. Addressing the High Power Computation requirements that are met while processing environmental data, our platform can be easily extended through tools that connect to GRID infrastructures, WCS web services, Bashyt API (for creating specialized hydrological reports) or any other specialized services (ex. graphics cluster visualization) that can be reached over the Internet. At run time, on the trainee's computer each tool is launched in an asynchronous running mode and connects to the data source that has been established by the teacher, retrieving and displaying the information to the user. The data transfer is accomplished directly between the trainee's computer and the corresponding services (e.g. OGC, Bashyt API, etc.) without passing through the core server platform. In this manner, the eGLE application can provide better and more responsive connections to a large number of users.

  2. OpenSearch (ECHO-ESIP) & REST API for Earth Science Data Access

    NASA Astrophysics Data System (ADS)

    Mitchell, A.; Cechini, M.; Pilone, D.

    2010-12-01

    This presentation will provide a brief technical overview of OpenSearch, the Earth Science Information Partners (ESIP) Federated Search framework, and the REST architecture; discuss NASA’s Earth Observing System (EOS) ClearingHOuse’s (ECHO) implementation lessons learned; and demonstrate the simplified usage of these technologies. SOAP, as a framework for web service communication has numerous advantages for Enterprise applications and Java/C# type programming languages. As a technical solution, SOAP has been a reliable framework on top of which many applications have been successfully developed and deployed. However, as interest grows for quick development cycles and more intriguing “mashups,” the SOAP API loses its appeal. Lightweight and simple are the vogue characteristics that are sought after. Enter the REST API architecture and OpenSearch format. Both of these items provide a new path for application development addressing some of the issues unresolved by SOAP. ECHO has made available all of its discovery, order submission, and data management services through a publicly accessible SOAP API. This interface is utilized by a variety of ECHO client and data partners to provide valuable capabilities to end users. As ECHO interacted with current and potential partners looking to develop Earth Science tools utilizing ECHO, it became apparent that the development overhead required to interact with the SOAP API was a growing barrier to entry. ECHO acknowledged the technical issues that were being uncovered by its partner community and chose to provide two new interfaces for interacting with the ECHO metadata catalog. The first interface is built upon the OpenSearch format and ESIP Federated Search framework. Leveraging these two items, a client (ECHO-ESIP) was developed with a focus on simplified searching and results presentation. The second interface is built upon the Representational State Transfer (REST) architecture. Leveraging the REST architecture, a new API has been made available that will provide access to the entire SOAP API suite of services. The results of these development activities has not only positioned to engage in the thriving world of mashup applications, but also provided an excellent real-world case study of how to successfully leverage these emerging technologies.

  3. CALIBRATION, OPTIMIZATION, AND SENSITIVITY AND UNCERTAINTY ALGORITHMS APPLICATION PROGRAMMING INTERFACE (COSU-API)

    EPA Science Inventory

    The Application Programming Interface (API) for Uncertainty Analysis, Sensitivity Analysis, and Parameter Estimation (UA/SA/PE API) tool development, here fore referred to as the Calibration, Optimization, and Sensitivity and Uncertainty Algorithms API (COSU-API), was initially d...

  4. The Ensembl REST API: Ensembl Data for Any Language.

    PubMed

    Yates, Andrew; Beal, Kathryn; Keenan, Stephen; McLaren, William; Pignatelli, Miguel; Ritchie, Graham R S; Ruffier, Magali; Taylor, Kieron; Vullo, Alessandro; Flicek, Paul

    2015-01-01

    We present a Web service to access Ensembl data using Representational State Transfer (REST). The Ensembl REST server enables the easy retrieval of a wide range of Ensembl data by most programming languages, using standard formats such as JSON and FASTA while minimizing client work. We also introduce bindings to the popular Ensembl Variant Effect Predictor tool permitting large-scale programmatic variant analysis independent of any specific programming language. The Ensembl REST API can be accessed at http://rest.ensembl.org and source code is freely available under an Apache 2.0 license from http://github.com/Ensembl/ensembl-rest. © The Author 2014. Published by Oxford University Press.

  5. Restful API Architecture Based on Laravel Framework

    NASA Astrophysics Data System (ADS)

    Chen, Xianjun; Ji, Zhoupeng; Fan, Yu; Zhan, Yongsong

    2017-10-01

    Web service has been an industry standard tech for message communication and integration between heterogeneous systems. RESTFUL API has become mainstream web service development paradigm after SOAP, how to effectively construct RESTFUL API remains a research hotspots. This paper presents a development model of RESTFUL API construction based on PHP language and LARAVEL framework. The key technical problems that need to be solved during the construction of RESTFUL API are discussed, and implementation details based on LARAVEL are given.

  6. Modular and scalable RESTful API to sustain STAR collaboration's record keeping

    NASA Astrophysics Data System (ADS)

    Arkhipkin, D.; Lauret, J.; Shanmuganathan, P. V.

    2015-12-01

    STAR collaboration's record system is a collection of heterogeneous and sparse information associated to each members and institutions. In its original incarnation, only flat information was stored revealing many restrictions such as the lack of historical change information, the inability to keep track of members leaving and re-joining STAR, or the ability to easily extend the saved information as new requirements appeared. In mid-2013, a new project was launched covering an extensive set of revisited requirements. The requirements led us to a design based on a RESTful API, back-end storage engine relying on key/value pair data representation model coupled with a tiered architecture design. This design was motivated by the fact that unifying many STAR tools, relying on the same business logic and storage engine, was a key and central feature for the maintainability and presentation of records. This central service API would leave no ambiguities and provide easy service integration between STAR tools. The new design stores the changes in records dynamically and allows tracking the changes chronologically. The storage engine is extensible as new field of information emerges (member specific or general) without affecting the presentation or the business logic layers. The new record system features a convenient administrative interface, fuzzy algorithms for data entry and search, and provides basic statistics and graphs. Finally, this modular approach is supplemented with access control, allowing private information and administrative operations to be hidden away from public eyes.

  7. Multi-National Information Sharing -- Cross Domain Collaborative Information Environment (CDCIE) Solution. Revision 4

    DTIC Science & Technology

    2005-04-12

    Hardware, Database, and Operating System independence using Java • Enterprise-class Architecture using Java2 Enterprise Edition 1.4 • Standards based...portal applications. Compliance with the Java Specification Request for Portlet APIs (JSR-168) (Portlet API) and Web Services for Remote Portals...authentication and authorization • Portal Standards using Java Specification Request for Portlet APIs (JSR-168) (Portlet API) and Web Services for Remote

  8. Registered File Support for Critical Operations Files at (Space Infrared Telescope Facility) SIRTF

    NASA Technical Reports Server (NTRS)

    Turek, G.; Handley, Tom; Jacobson, J.; Rector, J.

    2001-01-01

    The SIRTF Science Center's (SSC) Science Operations System (SOS) has to contend with nearly one hundred critical operations files via comprehensive file management services. The management is accomplished via the registered file system (otherwise known as TFS) which manages these files in a registered file repository composed of a virtual file system accessible via a TFS server and a file registration database. The TFS server provides controlled, reliable, and secure file transfer and storage by registering all file transactions and meta-data in the file registration database. An API is provided for application programs to communicate with TFS servers and the repository. A command line client implementing this API has been developed as a client tool. This paper describes the architecture, current implementation, but more importantly, the evolution of these services based on evolving community use cases and emerging information system technology.

  9. The Climate Data Analytic Services (CDAS) Framework.

    NASA Astrophysics Data System (ADS)

    Maxwell, T. P.; Duffy, D.

    2016-12-01

    Faced with unprecedented growth in climate data volume and demand, NASA has developed the Climate Data Analytic Services (CDAS) framework. This framework enables scientists to execute data processing workflows combining common analysis operations in a high performance environment close to the massive data stores at NASA. The data is accessed in standard (NetCDF, HDF, etc.) formats in a POSIX file system and processed using vetted climate data analysis tools (ESMF, CDAT, NCO, etc.). A dynamic caching architecture enables interactive response times. CDAS utilizes Apache Spark for parallelization and a custom array framework for processing huge datasets within limited memory spaces. CDAS services are accessed via a WPS API being developed in collaboration with the ESGF Compute Working Team to support server-side analytics for ESGF. The API can be accessed using either direct web service calls, a python script, a unix-like shell client, or a javascript-based web application. Client packages in python, scala, or javascript contain everything needed to make CDAS requests. The CDAS architecture brings together the tools, data storage, and high-performance computing required for timely analysis of large-scale data sets, where the data resides, to ultimately produce societal benefits. It is is currently deployed at NASA in support of the Collaborative REAnalysis Technical Environment (CREATE) project, which centralizes numerous global reanalysis datasets onto a single advanced data analytics platform. This service permits decision makers to investigate climate changes around the globe, inspect model trends and variability, and compare multiple reanalysis datasets.

  10. Mental illness and principal physical diagnoses among Asian American and Pacific Islander users of emergency services.

    PubMed

    Chen, Huey Jen

    2005-12-01

    The stigma of mental illness is one of the factors that prevents Asian Americans/Pacific Islanders (APIs) from seeking formal mental health services. A somatic complaint is more acceptable in expressing psychiatric/emotional distress. Admission diagnoses in API emergency service users with secondary psychiatric diagnoses were identified from the 2001 National Inpatient Sample (NIS) of the Healthcare Cost and Utilization Project (HCUP). The sample consisted of 10,623 adult APIs. The study examined the differences in the six leading principal physical admission diagnoses between API emergency service users with psychiatric diagnoses and those without psychiatric diagnoses. Several of the study findings create concern (e.g., the higher percentage of APIs with psychiatric diagnosis who were discharged against medical advice, the high percentage admitted with medication intoxication). Further study is needed to provide guidance for clinical practice.

  11. Preparing WIND for the STEREO Mission

    NASA Astrophysics Data System (ADS)

    Schroeder, P.; Ogilve, K.; Szabo, A.; Lin, R.; Luhmann, J.

    2006-05-01

    The upcoming STEREO mission's IMPACT and PLASTIC investigations will provide the first opportunity for long duration, detailed observations of 1 AU magnetic field structures, plasma ions and electrons, suprathermal electrons, and energetic particles at points bracketing Earth's heliospheric location. Stereoscopic/3D information from the STEREO SECCHI imagers and SWAVES radio experiment will make it possible to use both multipoint and quadrature studies to connect interplanetary Coronal Mass Ejections (ICME) and solar wind structures to CMEs and coronal holes observed at the Sun. To fully exploit these unique data sets, tight integration with similarly equipped missions at L1 will be essential, particularly WIND and ACE. The STEREO mission is building novel data analysis tools to take advantage of the mission's scientific potential. These tools will require reliable access and a well-documented interface to the L1 data sets. Such an interface already exists for ACE through the ACE Science Center. We plan to provide a similar service for the WIND mission that will supplement existing CDAWeb services. Building on tools also being developed for STEREO, we will create a SOAP application program interface (API) which will allow both our STEREO/WIND/ACE interactive browser and third-party software to access WIND data as a seamless and integral part of the STEREO mission. The API will also allow for more advanced forms of data mining than currently available through other data web services. Access will be provided to WIND-specific data analysis software as well. The development of cross-spacecraft data analysis tools will allow a larger scientific community to combine STEREO's unique in-situ data with those of other missions, particularly the L1 missions, and, therefore, to maximize STEREO's scientific potential in gaining a greater understanding of the heliosphere.

  12. Using USNO's API to Obtain Data

    NASA Astrophysics Data System (ADS)

    Lesniak, Michael V.; Pozniak, Daniel; Punnoose, Tarun

    2015-01-01

    The U.S. Naval Observatory (USNO) is in the process of modernizing its publicly available web services into APIs (Application Programming Interfaces). Services configured as APIs offer greater flexibility to the user and allow greater usage. Depending on the particular service, users who implement our APIs will receive either a PNG (Portable Network Graphics) image or data in JSON (JavaScript Object Notation) format. This raw data can then be embedded in third-party web sites or in apps.Part of the USNO's mission is to provide astronomical and timing data to government agencies and the general public. To this end, the USNO provides accurate computations of astronomical phenomena such as dates of lunar phases, rise and set times of the Moon and Sun, and lunar and solar eclipse times. Users who navigate to our web site and select one of our 18 services are prompted to complete a web form, specifying parameters such as date, time, location, and object. Many of our services work for years between 1700 and 2100, meaning that past, present, and future events can be computed. Upon form submission, our web server processes the request, computes the data, and outputs it to the user.Over recent years, the use of the web by the general public has vastly changed. In response to this, the USNO is modernizing its web-based data services. This includes making our computed data easier to embed within third-party web sites as well as more easily querying from apps running on tablets and smart phones. To facilitate this, the USNO has begun converting its services into APIs. In addition to the existing web forms for the various services, users are able to make direct URL requests that return either an image or numerical data.To date, four of our web services have been configured to run with APIs. Two are image-producing services: "Apparent Disk of a Solar System Object" and "Day and Night Across the Earth." Two API data services are "Complete Sun and Moon Data for One Day" and "Dates of Primary Phases of the Moon." Instructions for how to use our API services as well as examples of their use can be found on one of our explanatory web pages and will be discussed here.

  13. G2S: a web-service for annotating genomic variants on 3D protein structures.

    PubMed

    Wang, Juexin; Sheridan, Robert; Sumer, S Onur; Schultz, Nikolaus; Xu, Dong; Gao, Jianjiong

    2018-06-01

    Accurately mapping and annotating genomic locations on 3D protein structures is a key step in structure-based analysis of genomic variants detected by recent large-scale sequencing efforts. There are several mapping resources currently available, but none of them provides a web API (Application Programming Interface) that supports programmatic access. We present G2S, a real-time web API that provides automated mapping of genomic variants on 3D protein structures. G2S can align genomic locations of variants, protein locations, or protein sequences to protein structures and retrieve the mapped residues from structures. G2S API uses REST-inspired design and it can be used by various clients such as web browsers, command terminals, programming languages and other bioinformatics tools for bringing 3D structures into genomic variant analysis. The webserver and source codes are freely available at https://g2s.genomenexus.org. g2s@genomenexus.org. Supplementary data are available at Bioinformatics online.

  14. The Ensembl REST API: Ensembl Data for Any Language

    PubMed Central

    Yates, Andrew; Beal, Kathryn; Keenan, Stephen; McLaren, William; Pignatelli, Miguel; Ritchie, Graham R. S.; Ruffier, Magali; Taylor, Kieron; Vullo, Alessandro; Flicek, Paul

    2015-01-01

    Motivation: We present a Web service to access Ensembl data using Representational State Transfer (REST). The Ensembl REST server enables the easy retrieval of a wide range of Ensembl data by most programming languages, using standard formats such as JSON and FASTA while minimizing client work. We also introduce bindings to the popular Ensembl Variant Effect Predictor tool permitting large-scale programmatic variant analysis independent of any specific programming language. Availability and implementation: The Ensembl REST API can be accessed at http://rest.ensembl.org and source code is freely available under an Apache 2.0 license from http://github.com/Ensembl/ensembl-rest. Contact: ayates@ebi.ac.uk or flicek@ebi.ac.uk Supplementary information: Supplementary data are available at Bioinformatics online. PMID:25236461

  15. Integrated platform and API for electrophysiological data

    PubMed Central

    Sobolev, Andrey; Stoewer, Adrian; Leonhardt, Aljoscha; Rautenberg, Philipp L.; Kellner, Christian J.; Garbers, Christian; Wachtler, Thomas

    2014-01-01

    Recent advancements in technology and methodology have led to growing amounts of increasingly complex neuroscience data recorded from various species, modalities, and levels of study. The rapid data growth has made efficient data access and flexible, machine-readable data annotation a crucial requisite for neuroscientists. Clear and consistent annotation and organization of data is not only an important ingredient for reproducibility of results and re-use of data, but also essential for collaborative research and data sharing. In particular, efficient data management and interoperability requires a unified approach that integrates data and metadata and provides a common way of accessing this information. In this paper we describe GNData, a data management platform for neurophysiological data. GNData provides a storage system based on a data representation that is suitable to organize data and metadata from any electrophysiological experiment, with a functionality exposed via a common application programming interface (API). Data representation and API structure are compatible with existing approaches for data and metadata representation in neurophysiology. The API implementation is based on the Representational State Transfer (REST) pattern, which enables data access integration in software applications and facilitates the development of tools that communicate with the service. Client libraries that interact with the API provide direct data access from computing environments like Matlab or Python, enabling integration of data management into the scientist's experimental or analysis routines. PMID:24795616

  16. Integrated platform and API for electrophysiological data.

    PubMed

    Sobolev, Andrey; Stoewer, Adrian; Leonhardt, Aljoscha; Rautenberg, Philipp L; Kellner, Christian J; Garbers, Christian; Wachtler, Thomas

    2014-01-01

    Recent advancements in technology and methodology have led to growing amounts of increasingly complex neuroscience data recorded from various species, modalities, and levels of study. The rapid data growth has made efficient data access and flexible, machine-readable data annotation a crucial requisite for neuroscientists. Clear and consistent annotation and organization of data is not only an important ingredient for reproducibility of results and re-use of data, but also essential for collaborative research and data sharing. In particular, efficient data management and interoperability requires a unified approach that integrates data and metadata and provides a common way of accessing this information. In this paper we describe GNData, a data management platform for neurophysiological data. GNData provides a storage system based on a data representation that is suitable to organize data and metadata from any electrophysiological experiment, with a functionality exposed via a common application programming interface (API). Data representation and API structure are compatible with existing approaches for data and metadata representation in neurophysiology. The API implementation is based on the Representational State Transfer (REST) pattern, which enables data access integration in software applications and facilitates the development of tools that communicate with the service. Client libraries that interact with the API provide direct data access from computing environments like Matlab or Python, enabling integration of data management into the scientist's experimental or analysis routines.

  17. 49 CFR 195.205 - Repair, alteration and reconstruction of aboveground breakout tanks that have been in service.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ...-refrigerated and tanks built to API Standard 650 or its predecessor Standard 12C, repair, alteration, and reconstruction must be in accordance with API Standard 653. (2) For tanks built to API Specification 12F or API..., examination, and material requirements of those respective standards. (3) For high pressure tanks built to API...

  18. 49 CFR 195.205 - Repair, alteration and reconstruction of aboveground breakout tanks that have been in service.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ...-refrigerated and tanks built to API Standard 650 or its predecessor Standard 12C, repair, alteration, and reconstruction must be in accordance with API Standard 653. (2) For tanks built to API Specification 12F or API..., examination, and material requirements of those respective standards. (3) For high pressure tanks built to API...

  19. 49 CFR 195.205 - Repair, alteration and reconstruction of aboveground breakout tanks that have been in service.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ...-refrigerated and tanks built to API Standard 650 or its predecessor Standard 12C, repair, alteration, and reconstruction must be in accordance with API Standard 653. (2) For tanks built to API Specification 12F or API..., examination, and material requirements of those respective standards. (3) For high pressure tanks built to API...

  20. 49 CFR 195.205 - Repair, alteration and reconstruction of aboveground breakout tanks that have been in service.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ...-refrigerated and tanks built to API Standard 650 or its predecessor Standard 12C, repair, alteration, and reconstruction must be in accordance with API Standard 653. (2) For tanks built to API Specification 12F or API..., examination, and material requirements of those respective standards. (3) For high pressure tanks built to API...

  1. Oceanographic data at your fingertips: the SOCIB App for smartphones

    NASA Astrophysics Data System (ADS)

    Lora, Sebastian; Sebastian, Kristian; Troupin, Charles; Pau Beltran, Joan; Frontera, Biel; Gómara, Sonia; Tintoré, Joaquín

    2015-04-01

    The Balearic Islands Coastal Ocean Observing and Forecasting System (SOCIB, http://www.socib.es), is a multi-platform Marine Research Infrastructure that generates data from nearshore to the open sea in the Western Mediterranean Sea. In line with SOCIB principles of discoverable, freely available and standardized data, an application (App) for smartphones has been designed, with the objective of providing an easy access to all the data managed by SOCIB in real-time: underwater gliders, drifters, profiling buoys, research vessel, HF Radar and numerical model outputs (hydrodynamics and waves). The Data Centre, responsible for the aquisition, processing and visualisation of all SOCIB data, developed a REpresentational State Transfer (REST) application programming interface (API) called "DataDiscovery" (http://apps.socib.es/DataDiscovery/). This API is made up of RESTful web services that provide information on : platforms, instruments, deployments of instruments. It also provides the data themselves. In this way, it is possible to integrate SOCIB data in third-party applications, developed either by the Data Center or externally. The existence of a single point for the data distribution not only allows for an efficient management but also makes easier the concepts and data access for external developers, who are not necessarily familiar with the concepts and tools related to oceanographic or atmospheric data. The SOCIB App for Android (https://play.google.com/store/apps/details?id=com.socib) uses that API as a "data backend", in such a way that it is straightforward to manage which information is shown by the application, without having to modify and upload it again. The only pieces of information that do not depend on the services are the App "Sections" and "Screens", but the content displayed in each of them is obtained through requests to the web services. The API is not used only for the smartphone app: presently, most of SOCIB applications for data visualisation and access rely on the API, for instance: corporative web, deployment Application (Dapp, http://apps.socib.es/dapp/), Sea Boards (http://seaboard.socib.es/).

  2. 49 CFR 195.432 - Inspection of in-service breakout tanks.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... low-pressure steel aboveground breakout tanks according to API Standard 653 (incorporated by reference... breakout tanks built to API Standard 2510 according to section 6 of API 510. (d) The intervals of...

  3. 49 CFR 195.432 - Inspection of in-service breakout tanks.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... low-pressure steel aboveground breakout tanks according to API Standard 653 (incorporated by reference... breakout tanks built to API Standard 2510 according to section 6 of API 510. (d) The intervals of...

  4. 49 CFR 195.432 - Inspection of in-service breakout tanks.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... low-pressure steel aboveground breakout tanks according to API Standard 653 (incorporated by reference... breakout tanks built to API Standard 2510 according to section 6 of API 510. (d) The intervals of...

  5. 49 CFR 195.432 - Inspection of in-service breakout tanks.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... low-pressure steel aboveground breakout tanks according to API Standard 653 (incorporated by reference... breakout tanks built to API Standard 2510 according to section 6 of API 510. (d) The intervals of...

  6. 49 CFR 195.432 - Inspection of in-service breakout tanks.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... low-pressure steel aboveground breakout tanks according to API Standard 653 (incorporated by reference... breakout tanks built to API Standard 2510 according to section 6 of API 510. (d) The intervals of...

  7. NSLS-II HIGH LEVEL APPLICATION INFRASTRUCTURE AND CLIENT API DESIGN

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Shen, G.; Yang; L.

    2011-03-28

    The beam commissioning software framework of NSLS-II project adopts a client/server based architecture to replace the more traditional monolithic high level application approach. It is an open structure platform, and we try to provide a narrow API set for client application. With this narrow API, existing applications developed in different language under different architecture could be ported to our platform with small modification. This paper describes system infrastructure design, client API and system integration, and latest progress. As a new 3rd generation synchrotron light source with ultra low emittance, there are new requirements and challenges to control and manipulate themore » beam. A use case study and a theoretical analysis have been performed to clarify requirements and challenges to the high level applications (HLA) software environment. To satisfy those requirements and challenges, adequate system architecture of the software framework is critical for beam commissioning, study and operation. The existing traditional approaches are self-consistent, and monolithic. Some of them have adopted a concept of middle layer to separate low level hardware processing from numerical algorithm computing, physics modelling, data manipulating, plotting, and error handling. However, none of the existing approaches can satisfy the requirement. A new design has been proposed by introducing service oriented architecture technology. The HLA is combination of tools for accelerator physicists and operators, which is same as traditional approach. In NSLS-II, they include monitoring applications and control routines. Scripting environment is very important for the later part of HLA and both parts are designed based on a common set of APIs. Physicists and operators are users of these APIs, while control system engineers and a few accelerator physicists are the developers of these APIs. With our Client/Server mode based approach, we leave how to retrieve information to the developers of APIs and how to use them to form a physics application to the users. For example, how the channels are related to magnet and what the current real-time setting of a magnet is in physics unit are the internals of APIs. Measuring chromaticities are the users of APIs. All the users of APIs are working with magnet and instrument names in a physics unit. The low level communications in current or voltage unit are minimized. In this paper, we discussed our recent progress of our infrastructure development, and client API.« less

  8. E-DECIDER Decision Support Gateway For Earthquake Disaster Response

    NASA Astrophysics Data System (ADS)

    Glasscoe, M. T.; Stough, T. M.; Parker, J. W.; Burl, M. C.; Donnellan, A.; Blom, R. G.; Pierce, M. E.; Wang, J.; Ma, Y.; Rundle, J. B.; Yoder, M. R.

    2013-12-01

    Earthquake Data Enhanced Cyber-Infrastructure for Disaster Evaluation and Response (E-DECIDER) is a NASA-funded project developing capabilities for decision-making utilizing remote sensing data and modeling software in order to provide decision support for earthquake disaster management and response. E-DECIDER incorporates earthquake forecasting methodology and geophysical modeling tools developed through NASA's QuakeSim project in order to produce standards-compliant map data products to aid in decision-making following an earthquake. Remote sensing and geodetic data, in conjunction with modeling and forecasting tools, help provide both long-term planning information for disaster management decision makers as well as short-term information following earthquake events (i.e. identifying areas where the greatest deformation and damage has occurred and emergency services may need to be focused). E-DECIDER utilizes a service-based GIS model for its cyber-infrastructure in order to produce standards-compliant products for different user types with multiple service protocols (such as KML, WMS, WFS, and WCS). The goal is to make complex GIS processing and domain-specific analysis tools more accessible to general users through software services as well as provide system sustainability through infrastructure services. The system comprises several components, which include: a GeoServer for thematic mapping and data distribution, a geospatial database for storage and spatial analysis, web service APIs, including simple-to-use REST APIs for complex GIS functionalities, and geoprocessing tools including python scripts to produce standards-compliant data products. These are then served to the E-DECIDER decision support gateway (http://e-decider.org), the E-DECIDER mobile interface, and to the Department of Homeland Security decision support middleware UICDS (Unified Incident Command and Decision Support). The E-DECIDER decision support gateway features a web interface that delivers map data products including deformation modeling results (slope change and strain magnitude) and aftershock forecasts, with remote sensing change detection results under development. These products are event triggered (from the USGS earthquake feed) and will be posted to event feeds on the E-DECIDER webpage and accessible via the mobile interface and UICDS. E-DECIDER also features a KML service that provides infrastructure information from the FEMA HAZUS database through UICDS and the mobile interface. The back-end GIS service architecture and front-end gateway components form a decision support system that is designed for ease-of-use and extensibility for end-users.

  9. A Fine-Grained API Link Prediction Approach Supporting CMDA Mashup Recommendation

    NASA Astrophysics Data System (ADS)

    Zhang, J.; Bao, Q.; Lee, T. J.; Ramachandran, R.; Lee, S.; Pan, L.; Gatlin, P. N.; Maskey, M.

    2017-12-01

    Service (API) discovery and recommendation is key to the wide spread of service oriented architecture and service oriented software engineering. Service recommendation typically relies on service linkage prediction calculated by the semantic distances (or similarities) among services based on their collection of inherent attributes. Given a specific context (mashup goal), however, different attributes may contribute differently to a service linkage. In this work, instead of training a model for all attributes as a whole, a novel approach is presented to simultaneously train separate models for individual attributes. Our contributions are summarized in three-fold. First is that we have developed a scalable attribute-level data model, featuring scalability and extensibility. We have extended Multiplicative Attribute Graph (MAG) model to represent node profiles featuring rich categorical attributes, while relaxing its constraint of requiring a priori knowledge of predefined attributes. LDA is leveraged to dynamically identify attributes based on attribute modeling, and multiple Gaussian fit is applied to find global optimal values. The second contribution is that we have seamlessly integrated the latent relationships between API attributes as well as observed network structure based on historical API usage data. Such a layered information model enables us to predict the probability of a link between two APIs based on their attribute link affinities carrying a variety of information including meta data, semantic data, historical usage data, as well as crowdsourcing user comments and annotations. The third contribution is that we have developed a finegrained context-aware mashup-API recommendation technique. On top of individual models trained for separate attributes, a dedicated layer is trained to represent the latent attribute distribution regarding mashup purpose, i.e., sensitivity of attributes to context. Thus, given the description of an intended mashup, the attributes sensitive to the goal will be identified, and corresponding attribute models will be exploited to compute the possibility of API linkages under the context. Such a layered model increases search accuracy.

  10. Commercial Building Energy Asset Score

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    This software (Asset Scoring Tool) is designed to help building owners and managers to gain insight into the as-built efficiency of their buildings. It is a web tool where users can enter their building information and obtain an asset score report. The asset score report consists of modeled building energy use (by end use and by fuel type), building systems (envelope, lighting, heating, cooling, service hot water) evaluations, and recommended energy efficiency measures. The intended users are building owners and operators who have limited knowledge of building energy efficiency. The scoring tool collects minimum building data (~20 data entries) frommore » users and build a full-scale energy model using the inference functionalities from Facility Energy Decision System (FEDS). The scoring tool runs real-time building energy simulation using EnergyPlus and performs life-cycle cost analysis using FEDS. An API is also under development to allow the third-party applications to exchange data with the web service of the scoring tool.« less

  11. An overview of the CellML API and its implementation

    PubMed Central

    2010-01-01

    Background CellML is an XML based language for representing mathematical models, in a machine-independent form which is suitable for their exchange between different authors, and for archival in a model repository. Allowing for the exchange and archival of models in a computer readable form is a key strategic goal in bioinformatics, because of the associated improvements in scientific record accuracy, the faster iterative process of scientific development, and the ability to combine models into large integrative models. However, for CellML models to be useful, tools which can process them correctly are needed. Due to some of the more complex features present in CellML models, such as imports, developing code ab initio to correctly process models can be an onerous task. For this reason, there is a clear and pressing need for an application programming interface (API), and a good implementation of that API, upon which tools can base their support for CellML. Results We developed an API which allows the information in CellML models to be retrieved and/or modified. We also developed a series of optional extension APIs, for tasks such as simplifying the handling of connections between variables, dealing with physical units, validating models, and translating models into different procedural languages. We have also provided a Free/Open Source implementation of this application programming interface, optimised to achieve good performance. Conclusions Tools have been developed using the API which are mature enough for widespread use. The API has the potential to accelerate the development of additional tools capable of processing CellML, and ultimately lead to an increased level of sharing of mathematical model descriptions. PMID:20377909

  12. An overview of the CellML API and its implementation.

    PubMed

    Miller, Andrew K; Marsh, Justin; Reeve, Adam; Garny, Alan; Britten, Randall; Halstead, Matt; Cooper, Jonathan; Nickerson, David P; Nielsen, Poul F

    2010-04-08

    CellML is an XML based language for representing mathematical models, in a machine-independent form which is suitable for their exchange between different authors, and for archival in a model repository. Allowing for the exchange and archival of models in a computer readable form is a key strategic goal in bioinformatics, because of the associated improvements in scientific record accuracy, the faster iterative process of scientific development, and the ability to combine models into large integrative models.However, for CellML models to be useful, tools which can process them correctly are needed. Due to some of the more complex features present in CellML models, such as imports, developing code ab initio to correctly process models can be an onerous task. For this reason, there is a clear and pressing need for an application programming interface (API), and a good implementation of that API, upon which tools can base their support for CellML. We developed an API which allows the information in CellML models to be retrieved and/or modified. We also developed a series of optional extension APIs, for tasks such as simplifying the handling of connections between variables, dealing with physical units, validating models, and translating models into different procedural languages.We have also provided a Free/Open Source implementation of this application programming interface, optimised to achieve good performance. Tools have been developed using the API which are mature enough for widespread use. The API has the potential to accelerate the development of additional tools capable of processing CellML, and ultimately lead to an increased level of sharing of mathematical model descriptions.

  13. Phylesystem: a git-based data store for community-curated phylogenetic estimates.

    PubMed

    McTavish, Emily Jane; Hinchliff, Cody E; Allman, James F; Brown, Joseph W; Cranston, Karen A; Holder, Mark T; Rees, Jonathan A; Smith, Stephen A

    2015-09-01

    Phylogenetic estimates from published studies can be archived using general platforms like Dryad (Vision, 2010) or TreeBASE (Sanderson et al., 1994). Such services fulfill a crucial role in ensuring transparency and reproducibility in phylogenetic research. However, digital tree data files often require some editing (e.g. rerooting) to improve the accuracy and reusability of the phylogenetic statements. Furthermore, establishing the mapping between tip labels used in a tree and taxa in a single common taxonomy dramatically improves the ability of other researchers to reuse phylogenetic estimates. As the process of curating a published phylogenetic estimate is not error-free, retaining a full record of the provenance of edits to a tree is crucial for openness, allowing editors to receive credit for their work and making errors introduced during curation easier to correct. Here, we report the development of software infrastructure to support the open curation of phylogenetic data by the community of biologists. The backend of the system provides an interface for the standard database operations of creating, reading, updating and deleting records by making commits to a git repository. The record of the history of edits to a tree is preserved by git's version control features. Hosting this data store on GitHub (http://github.com/) provides open access to the data store using tools familiar to many developers. We have deployed a server running the 'phylesystem-api', which wraps the interactions with git and GitHub. The Open Tree of Life project has also developed and deployed a JavaScript application that uses the phylesystem-api and other web services to enable input and curation of published phylogenetic statements. Source code for the web service layer is available at https://github.com/OpenTreeOfLife/phylesystem-api. The data store can be cloned from: https://github.com/OpenTreeOfLife/phylesystem. A web application that uses the phylesystem web services is deployed at http://tree.opentreeoflife.org/curator. Code for that tool is available from https://github.com/OpenTreeOfLife/opentree. mtholder@gmail.com. © The Author 2015. Published by Oxford University Press.

  14. Dynamic Policy-Driven Quality of Service in Service-Oriented Information Management Systems

    DTIC Science & Technology

    2011-01-01

    both DiffServ and IntServ net- work QoS mechanisms. Wang et al [48] provide middleware APIs to shield applications from directly interacting with...complex network QoS mechanism APIs . Middleware frameworks transparently converted the specified application QoS requirements into low- er-level network...QoS mechanism APIs and provided network QoS assurances. Deployment-time resource allocation. Other prior work has focused on deploying ap- plications

  15. MSeqDR mvTool: A mitochondrial DNA Web and API resource for comprehensive variant annotation, universal nomenclature collation, and reference genome conversion.

    PubMed

    Shen, Lishuang; Attimonelli, Marcella; Bai, Renkui; Lott, Marie T; Wallace, Douglas C; Falk, Marni J; Gai, Xiaowu

    2018-06-01

    Accurate mitochondrial DNA (mtDNA) variant annotation is essential for the clinical diagnosis of diverse human diseases. Substantial challenges to this process include the inconsistency in mtDNA nomenclatures, the existence of multiple reference genomes, and a lack of reference population frequency data. Clinicians need a simple bioinformatics tool that is user-friendly, and bioinformaticians need a powerful informatics resource for programmatic usage. Here, we report the development and functionality of the MSeqDR mtDNA Variant Tool set (mvTool), a one-stop mtDNA variant annotation and analysis Web service. mvTool is built upon the MSeqDR infrastructure (https://mseqdr.org), with contributions of expert curated data from MITOMAP (https://www.mitomap.org) and HmtDB (https://www.hmtdb.uniba.it/hmdb). mvTool supports all mtDNA nomenclatures, converts variants to standard rCRS- and HGVS-based nomenclatures, and annotates novel mtDNA variants. Besides generic annotations from dbNSFP and Variant Effect Predictor (VEP), mvTool provides allele frequencies in more than 47,000 germline mitogenomes, and disease and pathogenicity classifications from MSeqDR, Mitomap, HmtDB and ClinVar (Landrum et al., 2013). mvTools also provides mtDNA somatic variants annotations. "mvTool API" is implemented for programmatic access using inputs in VCF, HGVS, or classical mtDNA variant nomenclatures. The results are reported as hyperlinked html tables, JSON, Excel, and VCF formats. MSeqDR mvTool is freely accessible at https://mseqdr.org/mvtool.php. © 2018 Wiley Periodicals, Inc.

  16. The Earth Data Analytic Services (EDAS) Framework

    NASA Astrophysics Data System (ADS)

    Maxwell, T. P.; Duffy, D.

    2017-12-01

    Faced with unprecedented growth in earth data volume and demand, NASA has developed the Earth Data Analytic Services (EDAS) framework, a high performance big data analytics framework built on Apache Spark. This framework enables scientists to execute data processing workflows combining common analysis operations close to the massive data stores at NASA. The data is accessed in standard (NetCDF, HDF, etc.) formats in a POSIX file system and processed using vetted earth data analysis tools (ESMF, CDAT, NCO, etc.). EDAS utilizes a dynamic caching architecture, a custom distributed array framework, and a streaming parallel in-memory workflow for efficiently processing huge datasets within limited memory spaces with interactive response times. EDAS services are accessed via a WPS API being developed in collaboration with the ESGF Compute Working Team to support server-side analytics for ESGF. The API can be accessed using direct web service calls, a Python script, a Unix-like shell client, or a JavaScript-based web application. New analytic operations can be developed in Python, Java, or Scala (with support for other languages planned). Client packages in Python, Java/Scala, or JavaScript contain everything needed to build and submit EDAS requests. The EDAS architecture brings together the tools, data storage, and high-performance computing required for timely analysis of large-scale data sets, where the data resides, to ultimately produce societal benefits. It is is currently deployed at NASA in support of the Collaborative REAnalysis Technical Environment (CREATE) project, which centralizes numerous global reanalysis datasets onto a single advanced data analytics platform. This service enables decision makers to compare multiple reanalysis datasets and investigate trends, variability, and anomalies in earth system dynamics around the globe.

  17. Cloud-Based Computational Tools for Earth Science Applications

    NASA Astrophysics Data System (ADS)

    Arendt, A. A.; Fatland, R.; Howe, B.

    2015-12-01

    Earth scientists are increasingly required to think across disciplines and utilize a wide range of datasets in order to solve complex environmental challenges. Although significant progress has been made in distributing data, researchers must still invest heavily in developing computational tools to accommodate their specific domain. Here we document our development of lightweight computational data systems aimed at enabling rapid data distribution, analytics and problem solving tools for Earth science applications. Our goal is for these systems to be easily deployable, scalable and flexible to accommodate new research directions. As an example we describe "Ice2Ocean", a software system aimed at predicting runoff from snow and ice in the Gulf of Alaska region. Our backend components include relational database software to handle tabular and vector datasets, Python tools (NumPy, pandas and xray) for rapid querying of gridded climate data, and an energy and mass balance hydrological simulation model (SnowModel). These components are hosted in a cloud environment for direct access across research teams, and can also be accessed via API web services using a REST interface. This API is a vital component of our system architecture, as it enables quick integration of our analytical tools across disciplines, and can be accessed by any existing data distribution centers. We will showcase several data integration and visualization examples to illustrate how our system has expanded our ability to conduct cross-disciplinary research.

  18. HMMER web server: 2018 update.

    PubMed

    Potter, Simon C; Luciani, Aurélien; Eddy, Sean R; Park, Youngmi; Lopez, Rodrigo; Finn, Robert D

    2018-06-14

    The HMMER webserver [http://www.ebi.ac.uk/Tools/hmmer] is a free-to-use service which provides fast searches against widely used sequence databases and profile hidden Markov model (HMM) libraries using the HMMER software suite (http://hmmer.org). The results of a sequence search may be summarized in a number of ways, allowing users to view and filter the significant hits by domain architecture or taxonomy. For large scale usage, we provide an application programmatic interface (API) which has been expanded in scope, such that all result presentations are available via both HTML and API. Furthermore, we have refactored our JavaScript visualization library to provide standalone components for different result representations. These consume the aforementioned API and can be integrated into third-party websites. The range of databases that can be searched against has been expanded, adding four sequence datasets (12 in total) and one profile HMM library (6 in total). To help users explore the biological context of their results, and to discover new data resources, search results are now supplemented with cross references to other EMBL-EBI databases.

  19. Internet SCADA Utilizing API's as Data Source

    NASA Astrophysics Data System (ADS)

    Robles, Rosslin John; Kim, Haeng-Kon; Kim, Tai-Hoon

    An Application programming interface or API is an interface implemented by a software program that enables it to interact with other software. Many companies provide free API services which can be utilized in Control Systems. SCADA is an example of a control system and it is a system that collects data from various sensors at a factory, plant or in other remote locations and then sends this data to a central computer which then manages and controls the data. In this paper, we designed a scheme for Weather Condition in Internet SCADA Environment utilizing data from external API services. The scheme was designed to double check the weather information in SCADA.

  20. The JANA calibrations and conditions database API

    NASA Astrophysics Data System (ADS)

    Lawrence, David

    2010-04-01

    Calibrations and conditions databases can be accessed from within the JANA Event Processing framework through the API defined in its JCalibration base class. The API is designed to support everything from databases, to web services to flat files for the backend. A Web Service backend using the gSOAP toolkit has been implemented which is particularly interesting since it addresses many modern cybersecurity issues including support for SSL. The API allows constants to be retrieved through a single line of C++ code with most of the context, including the transport mechanism, being implied by the run currently being analyzed and the environment relieving developers from implementing such details.

  1. The Proteins API: accessing key integrated protein and genome information

    PubMed Central

    Antunes, Ricardo; Alpi, Emanuele; Gonzales, Leonardo; Liu, Wudong; Luo, Jie; Qi, Guoying; Turner, Edd

    2017-01-01

    Abstract The Proteins API provides searching and programmatic access to protein and associated genomics data such as curated protein sequence positional annotations from UniProtKB, as well as mapped variation and proteomics data from large scale data sources (LSS). Using the coordinates service, researchers are able to retrieve the genomic sequence coordinates for proteins in UniProtKB. This, the LSS genomics and proteomics data for UniProt proteins is programmatically only available through this service. A Swagger UI has been implemented to provide documentation, an interface for users, with little or no programming experience, to ‘talk’ to the services to quickly and easily formulate queries with the services and obtain dynamically generated source code for popular programming languages, such as Java, Perl, Python and Ruby. Search results are returned as standard JSON, XML or GFF data objects. The Proteins API is a scalable, reliable, fast, easy to use RESTful services that provides a broad protein information resource for users to ask questions based upon their field of expertise and allowing them to gain an integrated overview of protein annotations available to aid their knowledge gain on proteins in biological processes. The Proteins API is available at (http://www.ebi.ac.uk/proteins/api/doc). PMID:28383659

  2. The Proteins API: accessing key integrated protein and genome information.

    PubMed

    Nightingale, Andrew; Antunes, Ricardo; Alpi, Emanuele; Bursteinas, Borisas; Gonzales, Leonardo; Liu, Wudong; Luo, Jie; Qi, Guoying; Turner, Edd; Martin, Maria

    2017-07-03

    The Proteins API provides searching and programmatic access to protein and associated genomics data such as curated protein sequence positional annotations from UniProtKB, as well as mapped variation and proteomics data from large scale data sources (LSS). Using the coordinates service, researchers are able to retrieve the genomic sequence coordinates for proteins in UniProtKB. This, the LSS genomics and proteomics data for UniProt proteins is programmatically only available through this service. A Swagger UI has been implemented to provide documentation, an interface for users, with little or no programming experience, to 'talk' to the services to quickly and easily formulate queries with the services and obtain dynamically generated source code for popular programming languages, such as Java, Perl, Python and Ruby. Search results are returned as standard JSON, XML or GFF data objects. The Proteins API is a scalable, reliable, fast, easy to use RESTful services that provides a broad protein information resource for users to ask questions based upon their field of expertise and allowing them to gain an integrated overview of protein annotations available to aid their knowledge gain on proteins in biological processes. The Proteins API is available at (http://www.ebi.ac.uk/proteins/api/doc). © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  3. Achieving Better Buying Power for Mobile Open Architecture Software Systems Through Diverse Acquisition Scenarios

    DTIC Science & Technology

    2016-04-30

    software (OSS) and proprietary (CSS) software elements or remote services (Scacchi, 2002, 2010), eventually including recent efforts to support Web ...specific platforms, including those operating on secured Web /mobile devices.  Common Development Technology provides AC development tools and common...transition to OA systems and OSS software elements, specifically for Web and Mobile devices within the realm of C3CB. OA, Open APIs, OSS, and CSS OA

  4. A Monte Carlo modeling alternative for the API Gamma Ray Calibration Facility.

    PubMed

    Galford, J E

    2017-04-01

    The gamma ray pit at the API Calibration Facility, located on the University of Houston campus, defines the API unit for natural gamma ray logs used throughout the petroleum logging industry. Future use of the facility is uncertain. An alternative method is proposed to preserve the gamma ray API unit definition as an industry standard by using Monte Carlo modeling to obtain accurate counting rate-to-API unit conversion factors for gross-counting and spectral gamma ray tool designs. Copyright © 2017 Elsevier Ltd. All rights reserved.

  5. Migrating Department of Defense (DoD) Web Service Based Applications to Mobile Computing Platforms

    DTIC Science & Technology

    2012-03-01

    World Wide Web Consortium (W3C) Geolocation API to identify the device’s location and then center the map on the device. Finally, we modify the entry...THIS PAGE INTENTIONALLY LEFT BLANK xii List of Acronyms and Abbreviations API Application Programming Interface CSS Cascading Style Sheets CLIMO...Java API for XML Web Services Reference Implementation JS JavaScript JSNI JavaScript Native Interface METOC Meteorological and Oceanographic MAA Mobile

  6. Physician visits and preventive care among Asian American and Pacific Islander long-term survivors of colorectal cancer, USA, 1996-2006.

    PubMed

    Steele, C Brooke; Townsend, Julie S; Tai, Eric; Thomas, Cheryll C

    2014-03-01

    Published literature on receipt of preventive healthcare services among Asian American and Pacific Islander (API) cancer survivors is scarce. We describe patterns in receipt of preventive services among API long-term colorectal cancer (CRC) survivors. Surveillance, Epidemiology, and End Results registry-Medicare data were used to identify 9,737 API and white patients who were diagnosed with CRC during 1996-2000 and who survived 5 or more years beyond their diagnoses. We examined receipt of vaccines, mammography (females), bone densitometry (females), and cholesterol screening among the survivors and how the physician specialties they visited for follow-up care correlated to services received. APIs were less likely than whites to receive mammography (52.0 vs. 69.3 %, respectively; P < 0.0001) but more likely to receive influenza vaccine, cholesterol screening, and bone densitometry. These findings remained significant in our multivariable model, except for receipt of bone densitometry. APIs visited PCPs only and both PCPs and oncologists more frequently than whites (P < 0.0001). Women who visited both PCPs and oncologists compared with PCPs only were more likely to receive mammography (odds ratio = 1.40; 95 % confidence interval, 1.05-1.86). Visits to both PCPs and oncologists were associated with increased use of mammography. Although API survivors visited these specialties more frequently than white survivors, API women may need culturally appropriate outreach to increase their use of this test. Long-term cancer survivors need to be aware of recommended preventive healthcare services, as well as who will manage their primary care and cancer surveillance follow-up.

  7. Multi-Robot Systems in Military Domains (Les Systemes Multi-Robots Dans les Domaines Militaires)

    DTIC Science & Technology

    2008-12-01

    to allow him to react quickly to improve his personal safety , it is mandatory to shorten the current very long delay needed for the human operator to...Hard RT tasks 2 OS / API Process monitoring 3 H / API Flexible communication medium 4 H / API Networking capabilities 5 H / API Safety 6 API...also be considered between high level services and legacy systems. 4) This is the one of the basic requirement for CoRoDe. 5) Safety : CRC, Timeouts

  8. The NetLogger Toolkit V2.0

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gunter, Dan; Lee, Jason; Stoufer, Martin

    2003-03-28

    The NetLogger Toolkit is designed to monitor, under actual operating conditions, the behavior of all the elements of the application-to-application communication path in order to determine exactly where time is spent within a complex system Using NetLogger, distnbuted application components are modified to produce timestamped logs of "interesting" events at all the critical points of the distributed system Events from each component are correlated, which allov^ one to characterize the performance of all aspects of the system and network in detail. The NetLogger Toolkit itself consists of four components an API and library of functions to simplify the generation ofmore » application-level event logs, a set of tools for collecting and sorting log files, an event archive system, and a tool for visualization and analysis of the log files In order to instrument an application to produce event logs, the application developer inserts calls to the NetLogger API at all the critical points in the code, then links the application with the NetLogger library All the tools in the NetLogger Toolkit share a common log format, and assume the existence of accurate and synchronized system clocks NetLogger messages can be logged using an easy-to-read text based format based on the lETF-proposed ULM format, or a binary format that can still be used through the same API but that is several times faster and smaller, with performance comparable or better than binary message formats such as MPI, XDR, SDDF-Binary, and PBIO. The NetLogger binary format is both highly efficient and self-describing, thus optimized for the dynamic message construction and parsing of application instrumentation. NetLogger includes an "activation" API that allows NetLogger logging to be turned on, off, or modified by changing an external file This IS useful for activating logging in daemons/services (e g GndFTP server). The NetLogger reliability API provides the ability to specify backup logging locations and penodically try to reconnect broken TCP pipe. A typical use for this is to store data on local disk while net is down. An event archiver can log one or more incoming NetLogger streams to a local disk file (netlogd) or to a mySQL database (netarchd). We have found exploratory, visual analysis of the log event data to be the most useful means of determining the causes of performance anomalies The NetLogger Visualization tool, niv, has been developed to provide a flexible and interactive graphical representation of system-level and application-level events.« less

  9. Introducing the PRIDE Archive RESTful web services.

    PubMed

    Reisinger, Florian; del-Toro, Noemi; Ternent, Tobias; Hermjakob, Henning; Vizcaíno, Juan Antonio

    2015-07-01

    The PRIDE (PRoteomics IDEntifications) database is one of the world-leading public repositories of mass spectrometry (MS)-based proteomics data and it is a founding member of the ProteomeXchange Consortium of proteomics resources. In the original PRIDE database system, users could access data programmatically by accessing the web services provided by the PRIDE BioMart interface. New REST (REpresentational State Transfer) web services have been developed to serve the most popular functionality provided by BioMart (now discontinued due to data scalability issues) and address the data access requirements of the newly developed PRIDE Archive. Using the API (Application Programming Interface) it is now possible to programmatically query for and retrieve peptide and protein identifications, project and assay metadata and the originally submitted files. Searching and filtering is also possible by metadata information, such as sample details (e.g. species and tissues), instrumentation (mass spectrometer), keywords and other provided annotations. The PRIDE Archive web services were first made available in April 2014. The API has already been adopted by a few applications and standalone tools such as PeptideShaker, PRIDE Inspector, the Unipept web application and the Python-based BioServices package. This application is free and open to all users with no login requirement and can be accessed at http://www.ebi.ac.uk/pride/ws/archive/. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.

  10. Using the RxNorm web services API for quality assurance purposes.

    PubMed

    Peters, Lee; Bodenreider, Olivier

    2008-11-06

    Auditing large, rapidly evolving terminological systems is still a challenge. In the case of RxNorm, a standardized nomenclature for clinical drugs, we argue that quality assurance processes can benefit from the recently released application programming interface (API) provided by RxNav. We demonstrate the usefulness of the API by performing a systematic comparison of alternative paths in the RxNorm graph, over several thousands of drug entities. This study revealed potential errors in RxNorm, currently under review. The results also prompted us to modify the implementation of RxNav to navigate the RxNorm graph more accurately. The RxNav web services API used in this experiment is robust and fast.

  11. Using the RxNorm Web Services API for Quality Assurance Purposes

    PubMed Central

    Peters, Lee; Bodenreider, Olivier

    2008-01-01

    Auditing large, rapidly evolving terminological systems is still a challenge. In the case of RxNorm, a standardized nomenclature for clinical drugs, we argue that quality assurance processes can benefit from the recently released application programming interface (API) provided by RxNav. We demonstrate the usefulness of the API by performing a systematic comparison of alternative paths in the RxNorm graph, over several thousands of drug entities. This study revealed potential errors in RxNorm, currently under review. The results also prompted us to modify the implementation of RxNav to navigate the RxNorm graph more accurately. The RxNorm web services API used in this experiment is robust and fast. PMID:18999038

  12. Transition from Legacy to Connectivity Solution for Infrastructure Control of Smart Municipal Systems

    NASA Astrophysics Data System (ADS)

    Zabasta, A.; Kunicina, N.; Kondratjevs, K.

    2017-06-01

    Collaboration between heterogeneous systems and architectures is not an easy problem in the automation domain. By now, utilities and suppliers encounter real problems due to underestimated costs of technical solutions, frustration in selecting technical solutions relevant for local needs, and incompatibilities between a plenty of protocols and appropriate solutions. The paper presents research on creation of architecture of smart municipal systems in a local cloud of services that apply SOA and IoT approaches. The authors of the paper have developed a broker that applies orchestration services and resides on a gateway, which provides adapter and protocol translation functions, as well as applies a tool for wiring together hardware devices, APIs and online services.

  13. CyanoBase: the cyanobacteria genome database update 2010.

    PubMed

    Nakao, Mitsuteru; Okamoto, Shinobu; Kohara, Mitsuyo; Fujishiro, Tsunakazu; Fujisawa, Takatomo; Sato, Shusei; Tabata, Satoshi; Kaneko, Takakazu; Nakamura, Yasukazu

    2010-01-01

    CyanoBase (http://genome.kazusa.or.jp/cyanobase) is the genome database for cyanobacteria, which are model organisms for photosynthesis. The database houses cyanobacteria species information, complete genome sequences, genome-scale experiment data, gene information, gene annotations and mutant information. In this version, we updated these datasets and improved the navigation and the visual display of the data views. In addition, a web service API now enables users to retrieve the data in various formats with other tools, seamlessly.

  14. Factor structure of the autonomy preference index in people with severe mental illness.

    PubMed

    Bonfils, Kelsey A; Adams, Erin L; Mueser, Kim T; Wright-Berryman, Jennifer L; Salyers, Michelle P

    2015-08-30

    People vary in the amount of control they want to exercise over decisions about their healthcare. Given the importance of patient-centered care, accurate measurement of these autonomy preferences is critical. This study aimed to assess the factor structure of the Autonomy Preference Index (API), used widely in general healthcare, in individuals with severe mental illness. Data came from two studies of people with severe mental illness (N=293) who were receiving mental health and/or primary care/integrated care services. Autonomy preferences were assessed with the API regarding both psychiatric and primary care services. Confirmatory factor analysis was used to evaluate fit of the hypothesized two-factor structure of the API (decision-making autonomy and information-seeking autonomy). Results indicated the hypothesized structure for the API did not adequately fit the data for either psychiatric or primary care services. Three problematic items were dropped, resulting in adequate fit for both types of treatment. These results suggest that with relatively minor modifications the API has an acceptable factor structure when asking people with severe mental illness about their preferences to be involved in decision-making. The modified API has clinical and research utility for this population in the burgeoning field of autonomy in patient-centered healthcare. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  15. A Driving Cycle Detection Approach Using Map Service API

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhu, Lei; Gonder, Jeffrey D

    Following advancements in smartphone and portable global positioning system (GPS) data collection, wearable GPS data have realized extensive use in transportation surveys and studies. The task of detecting driving cycles (driving or car-mode trajectory segments) from wearable GPS data has been the subject of much research. Specifically, distinguishing driving cycles from other motorized trips (such as taking a bus) is the main research problem in this paper. Many mode detection methods only focus on raw GPS speed data while some studies apply additional information, such as geographic information system (GIS) data, to obtain better detection performance. Procuring and maintaining dedicatedmore » road GIS data are costly and not trivial, whereas the technical maturity and broad use of map service application program interface (API) queries offers opportunities for mode detection tasks. The proposed driving cycle detection method takes advantage of map service APIs to obtain high-quality car-mode API route information and uses a trajectory segmentation algorithm to find the best-matched API route. The car-mode API route data combined with the actual route information, including the actual mode information, are used to train a logistic regression machine learning model, which estimates car modes and non-car modes with probability rates. The experimental results show promise for the proposed method's ability to detect vehicle mode accurately.« less

  16. Improving data discoverability, accessibility, and interoperability with the Esri ArcGIS Platform at the NASA Atmospheric Science Data Center (ASDC).

    NASA Astrophysics Data System (ADS)

    Tisdale, M.

    2017-12-01

    NASA's Atmospheric Science Data Center (ASDC) is operationally using the Esri ArcGIS Platform to improve data discoverability, accessibility and interoperability to meet the diversifying user requirements from government, private, public and academic communities. The ASDC is actively working to provide their mission essential datasets as ArcGIS Image Services, Open Geospatial Consortium (OGC) Web Mapping Services (WMS), and OGC Web Coverage Services (WCS) while leveraging the ArcGIS multidimensional mosaic dataset structure. Science teams at ASDC are utilizing these services through the development of applications using the Web AppBuilder for ArcGIS and the ArcGIS API for Javascript. These services provide greater exposure of ASDC data holdings to the GIS community and allow for broader sharing and distribution to various end users. These capabilities provide interactive visualization tools and improved geospatial analytical tools for a mission critical understanding in the areas of the earth's radiation budget, clouds, aerosols, and tropospheric chemistry. The presentation will cover how the ASDC is developing geospatial web services and applications to improve data discoverability, accessibility, and interoperability.

  17. An Auto-Configuration System for the GMSEC Architecture and API

    NASA Technical Reports Server (NTRS)

    Moholt, Joseph; Mayorga, Arturo

    2007-01-01

    A viewgraph presentation on an automated configuration concept for The Goddard Mission Services Evolution Center (GMSEC) architecture and Application Program Interface (API) is shown. The topics include: 1) The Goddard Mission Services Evolution Center (GMSEC); 2) Automated Configuration Concept; 3) Implementation Approach; and 4) Key Components and Benefits.

  18. Environmental Models as a Service: Enabling Interoperability ...

    EPA Pesticide Factsheets

    Achieving interoperability in environmental modeling has evolved as software technology has progressed. The recent rise of cloud computing and proliferation of web services initiated a new stage for creating interoperable systems. Scientific programmers increasingly take advantage of streamlined deployment processes and affordable cloud access to move algorithms and data to the web for discoverability and consumption. In these deployments, environmental models can become available to end users through RESTful web services and consistent application program interfaces (APIs) that consume, manipulate, and store modeling data. RESTful modeling APIs also promote discoverability and guide usability through self-documentation. Embracing the RESTful paradigm allows models to be accessible via a web standard, and the resulting endpoints are platform- and implementation-agnostic while simultaneously presenting significant computational capabilities for spatial and temporal scaling. RESTful APIs present data in a simple verb-noun web request interface: the verb dictates how a resource is consumed using HTTP methods (e.g., GET, POST, and PUT) and the noun represents the URL reference of the resource on which the verb will act. The RESTful API can self-document in both the HTTP response and an interactive web page using the Open API standard. This lets models function as an interoperable service that promotes sharing, documentation, and discoverability. Here, we discuss the

  19. Developer Network

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    2012-08-21

    NREL's Developer Network, developer.nrel.gov, provides data that users can access to provide data to their own analyses, mobile and web applications. Developers can retrieve the data through a Web services API (application programming interface). The Developer Network handles overhead of serving up web services such as key management, authentication, analytics, reporting, documentation standards, and throttling in a common architecture, while allowing web services and APIs to be maintained and managed independently.

  20. 49 CFR 195.307 - Pressure testing aboveground breakout tanks.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... aboveground breakout tanks. (a) For aboveground breakout tanks built into API Specification 12F and first placed in service after October 2, 2000, pneumatic testing must be in accordance with section 5.3 of API Specification 12 F (incorporated by reference, see § 195.3). (b) For aboveground breakout tanks built to API...

  1. 49 CFR 195.405 - Protection against ignitions and safe access/egress involving floating roofs.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... accordance with API Recommended Practice 2003, unless the operator notes in the procedural manual (§ 195.402(c)) why compliance with all or certain provisions of API Recommended Practice 2003 is not necessary... removed from service for cleaning) are addressed in API Publication 2026. After October 2, 2000, the...

  2. 49 CFR 195.405 - Protection against ignitions and safe access/egress involving floating roofs.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... accordance with API Recommended Practice 2003, unless the operator notes in the procedural manual (§ 195.402(c)) why compliance with all or certain provisions of API Recommended Practice 2003 is not necessary... removed from service for cleaning) are addressed in API Publication 2026. After October 2, 2000, the...

  3. 49 CFR 195.405 - Protection against ignitions and safe access/egress involving floating roofs.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... accordance with API Recommended Practice 2003, unless the operator notes in the procedural manual (§ 195.402(c)) why compliance with all or certain provisions of API Recommended Practice 2003 is not necessary... removed from service for cleaning) are addressed in API Publication 2026. After October 2, 2000, the...

  4. 49 CFR 195.405 - Protection against ignitions and safe access/egress involving floating roofs.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... accordance with API Recommended Practice 2003, unless the operator notes in the procedural manual (§ 195.402(c)) why compliance with all or certain provisions of API Recommended Practice 2003 is not necessary... removed from service for cleaning) are addressed in API Publication 2026. After October 2, 2000, the...

  5. 49 CFR 195.307 - Pressure testing aboveground breakout tanks.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... aboveground breakout tanks. (a) For aboveground breakout tanks built into API Specification 12F and first placed in service after October 2, 2000, pneumatic testing must be in accordance with section 5.3 of API Specification 12 F (incorporated by reference, see § 195.3). (b) For aboveground breakout tanks built to API...

  6. 49 CFR 195.405 - Protection against ignitions and safe access/egress involving floating roofs.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... accordance with API Recommended Practice 2003, unless the operator notes in the procedural manual (§ 195.402(c)) why compliance with all or certain provisions of API Recommended Practice 2003 is not necessary... removed from service for cleaning) are addressed in API Publication 2026. After October 2, 2000, the...

  7. 49 CFR 195.307 - Pressure testing aboveground breakout tanks.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... aboveground breakout tanks. (a) For aboveground breakout tanks built into API Specification 12F and first placed in service after October 2, 2000, pneumatic testing must be in accordance with section 5.3 of API Specification 12 F (incorporated by reference, see § 195.3). (b) For aboveground breakout tanks built to API...

  8. 49 CFR 195.307 - Pressure testing aboveground breakout tanks.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... aboveground breakout tanks. (a) For aboveground breakout tanks built into API Specification 12F and first placed in service after October 2, 2000, pneumatic testing must be in accordance with section 5.3 of API Specification 12 F (incorporated by reference, see § 195.3). (b) For aboveground breakout tanks built to API...

  9. 49 CFR 195.307 - Pressure testing aboveground breakout tanks.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... placed in service after October 2, 2000, pneumatic testing must be in accordance with section 5.3 of API Specification 12 F (incorporated by reference, see § 195.3). (b) For aboveground breakout tanks built to API Standard 620 and first placed in service after October 2, 2000, hydrostatic and pneumatic testing must be...

  10. A WPS Based Architecture for Climate Data Analytic Services (CDAS) at NASA

    NASA Astrophysics Data System (ADS)

    Maxwell, T. P.; McInerney, M.; Duffy, D.; Carriere, L.; Potter, G. L.; Doutriaux, C.

    2015-12-01

    Faced with unprecedented growth in the Big Data domain of climate science, NASA has developed the Climate Data Analytic Services (CDAS) framework. This framework enables scientists to execute trusted and tested analysis operations in a high performance environment close to the massive data stores at NASA. The data is accessed in standard (NetCDF, HDF, etc.) formats in a POSIX file system and processed using trusted climate data analysis tools (ESMF, CDAT, NCO, etc.). The framework is structured as a set of interacting modules allowing maximal flexibility in deployment choices. The current set of module managers include: Staging Manager: Runs the computation locally on the WPS server or remotely using tools such as celery or SLURM. Compute Engine Manager: Runs the computation serially or distributed over nodes using a parallelization framework such as celery or spark. Decomposition Manger: Manages strategies for distributing the data over nodes. Data Manager: Handles the import of domain data from long term storage and manages the in-memory and disk-based caching architectures. Kernel manager: A kernel is an encapsulated computational unit which executes a processor's compute task. Each kernel is implemented in python exploiting existing analysis packages (e.g. CDAT) and is compatible with all CDAS compute engines and decompositions. CDAS services are accessed via a WPS API being developed in collaboration with the ESGF Compute Working Team to support server-side analytics for ESGF. The API can be executed using either direct web service calls, a python script or application, or a javascript-based web application. Client packages in python or javascript contain everything needed to make CDAS requests. The CDAS architecture brings together the tools, data storage, and high-performance computing required for timely analysis of large-scale data sets, where the data resides, to ultimately produce societal benefits. It is is currently deployed at NASA in support of the Collaborative REAnalysis Technical Environment (CREATE) project, which centralizes numerous global reanalysis datasets onto a single advanced data analytics platform. This service permits decision makers to investigate climate changes around the globe, inspect model trends, compare multiple reanalysis datasets, and variability.

  11. CyanoBase: the cyanobacteria genome database update 2010

    PubMed Central

    Nakao, Mitsuteru; Okamoto, Shinobu; Kohara, Mitsuyo; Fujishiro, Tsunakazu; Fujisawa, Takatomo; Sato, Shusei; Tabata, Satoshi; Kaneko, Takakazu; Nakamura, Yasukazu

    2010-01-01

    CyanoBase (http://genome.kazusa.or.jp/cyanobase) is the genome database for cyanobacteria, which are model organisms for photosynthesis. The database houses cyanobacteria species information, complete genome sequences, genome-scale experiment data, gene information, gene annotations and mutant information. In this version, we updated these datasets and improved the navigation and the visual display of the data views. In addition, a web service API now enables users to retrieve the data in various formats with other tools, seamlessly. PMID:19880388

  12. FirebrowseR: an R client to the Broad Institute’s Firehose Pipeline

    PubMed Central

    Deng, Mario; Brägelmann, Johannes; Kryukov, Ivan; Saraiva-Agostinho, Nuno; Perner, Sven

    2017-01-01

    With its Firebrowse service (http://firebrowse.org/) the Broad Institute is making large-scale multi-platform omics data analysis results publicly available through a Representational State Transfer (REST) Application Programmable Interface (API). Querying this database through an API client from an arbitrary programming environment is an essential task, allowing other developers and researchers to focus on their analysis and avoid data wrangling. Hence, as a first result, we developed a workflow to automatically generate, test and deploy such clients for rapid response to API changes. Its underlying infrastructure, a combination of free and publicly available web services, facilitates the development of API clients. It decouples changes in server software from the client software by reacting to changes in the RESTful service and removing direct dependencies on a specific implementation of an API. As a second result, FirebrowseR, an R client to the Broad Institute’s RESTful Firehose Pipeline, is provided as a working example, which is built by the means of the presented workflow. The package’s features are demonstrated by an example analysis of cancer gene expression data. Database URL: https://github.com/mariodeng/ PMID:28062517

  13. FirebrowseR: an R client to the Broad Institute's Firehose Pipeline.

    PubMed

    Deng, Mario; Brägelmann, Johannes; Kryukov, Ivan; Saraiva-Agostinho, Nuno; Perner, Sven

    2017-01-01

    With its Firebrowse service (http://firebrowse.org/) the Broad Institute is making large-scale multi-platform omics data analysis results publicly available through a Representational State Transfer (REST) Application Programmable Interface (API). Querying this database through an API client from an arbitrary programming environment is an essential task, allowing other developers and researchers to focus on their analysis and avoid data wrangling. Hence, as a first result, we developed a workflow to automatically generate, test and deploy such clients for rapid response to API changes. Its underlying infrastructure, a combination of free and publicly available web services, facilitates the development of API clients. It decouples changes in server software from the client software by reacting to changes in the RESTful service and removing direct dependencies on a specific implementation of an API. As a second result, FirebrowseR, an R client to the Broad Institute's RESTful Firehose Pipeline, is provided as a working example, which is built by the means of the presented workflow. The package's features are demonstrated by an example analysis of cancer gene expression data.Database URL: https://github.com/mariodeng/. © The Author(s) 2017. Published by Oxford University Press.

  14. Using Map Service API for Driving Cycle Detection for Wearable GPS Data: Preprint

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhu, Lei; Gonder, Jeffrey D

    Following advancements in smartphone and portable global positioning system (GPS) data collection, wearable GPS data have realized extensive use in transportation surveys and studies. The task of detecting driving cycles (driving or car-mode trajectory segments) from wearable GPS data has been the subject of much research. Specifically, distinguishing driving cycles from other motorized trips (such as taking a bus) is the main research problem in this paper. Many mode detection methods only focus on raw GPS speed data while some studies apply additional information, such as geographic information system (GIS) data, to obtain better detection performance. Procuring and maintaining dedicatedmore » road GIS data are costly and not trivial, whereas the technical maturity and broad use of map service application program interface (API) queries offers opportunities for mode detection tasks. The proposed driving cycle detection method takes advantage of map service APIs to obtain high-quality car-mode API route information and uses a trajectory segmentation algorithm to find the best-matched API route. The car-mode API route data combined with the actual route information, including the actual mode information, are used to train a logistic regression machine learning model, which estimates car modes and non-car modes with probability rates. The experimental results show promise for the proposed method's ability to detect vehicle mode accurately.« less

  15. Data Services in Support of High Performance Computing-Based Distributed Hydrologic Models

    NASA Astrophysics Data System (ADS)

    Tarboton, D. G.; Horsburgh, J. S.; Dash, P. K.; Gichamo, T.; Yildirim, A. A.; Jones, N.

    2014-12-01

    We have developed web-based data services to support the application of hydrologic models on High Performance Computing (HPC) systems. The purposes of these services are to provide hydrologic researchers, modelers, water managers, and users access to HPC resources without requiring them to become HPC experts and understanding the intrinsic complexities of the data services, so as to reduce the amount of time and effort spent in finding and organizing the data required to execute hydrologic models and data preprocessing tools on HPC systems. These services address some of the data challenges faced by hydrologic models that strive to take advantage of HPC. Needed data is often not in the form needed by such models, requiring researchers to spend time and effort on data preparation and preprocessing that inhibits or limits the application of these models. Another limitation is the difficult to use batch job control and queuing systems used by HPC systems. We have developed a REST-based gateway application programming interface (API) for authenticated access to HPC systems that abstracts away many of the details that are barriers to HPC use and enhances accessibility from desktop programming and scripting languages such as Python and R. We have used this gateway API to establish software services that support the delineation of watersheds to define a modeling domain, then extract terrain and land use information to automatically configure the inputs required for hydrologic models. These services support the Terrain Analysis Using Digital Elevation Model (TauDEM) tools for watershed delineation and generation of hydrology-based terrain information such as wetness index and stream networks. These services also support the derivation of inputs for the Utah Energy Balance snowmelt model used to address questions such as how climate, land cover and land use change may affect snowmelt inputs to runoff generation. To enhance access to the time varying climate data used to drive hydrologic models, we have developed services to downscale and re-grid nationally available climate analysis data from systems such as NLDAS and MERRA. These cases serve as examples for how this approach can be extended to other models to enhance the use of HPC for hydrologic modeling.

  16. From Analysis to Impact: Challenges and Outcomes from Google's Cloud-based Platforms for Analyzing and Leveraging Petapixels of Geospatial Data

    NASA Astrophysics Data System (ADS)

    Thau, D.

    2017-12-01

    For the past seven years, Google has made petabytes of Earth observation data, and the tools to analyze it, freely available to researchers around the world via cloud computing. These data and tools were initially available via Google Earth Engine and are increasingly available on the Google Cloud Platform. We have introduced a number of APIs for both the analysis and presentation of geospatial data that have been successfully used to create impactful datasets and web applications, including studies of global surface water availability, global tree cover change, and crop yield estimation. Each of these projects used the cloud to analyze thousands to millions of Landsat scenes. The APIs support a range of publishing options, from outputting imagery and data for inclusion in papers, to providing tools for full scale web applications that provide analysis tools of their own. Over the course of developing these tools, we have learned a number of lessons about how to build a publicly available cloud platform for geospatial analysis, and about how the characteristics of an API can affect the kinds of impacts a platform can enable. This study will present an overview of how Google Earth Engine works and how Google's geospatial capabilities are extending to Google Cloud Platform. We will provide a number of case studies describing how these platforms, and the data they host, have been leveraged to build impactful decision support tools used by governments, researchers, and other institutions, and we will describe how the available APIs have shaped (or constrained) those tools. [Image Credit: Tyler A. Erickson

  17. CMR Catalog Service for the Web

    NASA Technical Reports Server (NTRS)

    Newman, Doug; Mitchell, Andrew

    2016-01-01

    With the impending retirement of Global Change Master Directory (GCMD) Application Programming Interfaces (APIs) the Common Metadata Repository (CMR) was charged with providing a collection-level Catalog Service for the Web (CSW) that provided the same level of functionality as GCMD. This talk describes the capabilities of the CMR CSW API with particular reference to the support of the Committee on Earth Observation Satellites (CEOS) Working Group on Information Systems and Services (WGISS) Integrated Catalog (CWIC).

  18. Commercial Building Energy Saver, API

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hong, Tianzhen; Piette, Mary; Lee, Sang Hoon

    2015-08-27

    The CBES API provides Application Programming Interface to a suite of functions to improve energy efficiency of buildings, including building energy benchmarking, preliminary retrofit analysis using a pre-simulation database DEEP, and detailed retrofit analysis using energy modeling with the EnergyPlus simulation engine. The CBES API is used to power the LBNL CBES Web App. It can be adopted by third party developers and vendors into their software tools and platforms.

  19. Interactive Learning Modules: Enabling Near Real-Time Oceanographic Data Use In Undergraduate Education

    NASA Astrophysics Data System (ADS)

    Kilb, D. L.; Fundis, A. T.; Risien, C. M.

    2012-12-01

    The focus of the Education and Public Engagement (EPE) component of the NSF's Ocean Observatories Initiative (OOI) is to provide a new layer of cyber-interactivity for undergraduate educators to bring near real-time data from the global ocean into learning environments. To accomplish this, we are designing six online services including: 1) visualization tools, 2) a lesson builder, 3) a concept map builder, 4) educational web services (middleware), 5) collaboration tools and 6) an educational resource database. Here, we report on our Fall 2012 release that includes the first four of these services: 1) Interactive visualization tools allow users to interactively select data of interest, display the data in various views (e.g., maps, time-series and scatter plots) and obtain statistical measures such as mean, standard deviation and a regression line fit to select data. Specific visualization tools include a tool to compare different months of data, a time series explorer tool to investigate the temporal evolution of select data parameters (e.g., sea water temperature or salinity), a glider profile tool that displays ocean glider tracks and associated transects, and a data comparison tool that allows users to view the data either in scatter plot view comparing one parameter with another, or in time series view. 2) Our interactive lesson builder tool allows users to develop a library of online lesson units, which are collaboratively editable and sharable and provides starter templates designed from learning theory knowledge. 3) Our interactive concept map tool allows the user to build and use concept maps, a graphical interface to map the connection between concepts and ideas. This tool also provides semantic-based recommendations, and allows for embedding of associated resources such as movies, images and blogs. 4) Education web services (middleware) will provide an educational resource database API.

  20. [Interview Questions

    NASA Technical Reports Server (NTRS)

    Smith, Dan

    2007-01-01

    The Goddard Mission Services Evolution Center, or GMSEC, was started in 2001 to create a new standard approach for managing GSFC missions. Standardized approaches in the past involved selecting and then integrating the most appropriate set of functional tools. Assumptions were made that "one size fits all" and that tool changes would not be necessary for many years. GMSEC took a very different approach and has proven to be very successful. The core of the GMSEC architecture consists of a publish/subscribe message bus, standardized message formats, and an Applications Programming Interface (API). The API supports multiple operating systems, programming languages and messaging middleware products. We use a GMSEC-developed free middleware for low-cost development. A high capacity, robust middleware is used for operations and a messaging system with a very small memory footprint is used for on-board flight software. Software components can use the standard message formats or develop adapters to convert from their native formats to the GMSEC formats. We do not want vendors to modify their core products. Over 50 software components are now available for use with the GMSEC architecture. Most available commercial telemetry and command systems, including the GMV hifly Satellite Control System, have been adapted to run in the GMSEC labs.

  1. Web GIS in practice III: creating a simple interactive map of England's Strategic Health Authorities using Google Maps API, Google Earth KML, and MSN Virtual Earth Map Control

    PubMed Central

    Boulos, Maged N Kamel

    2005-01-01

    This eye-opener article aims at introducing the health GIS community to the emerging online consumer geoinformatics services from Google and Microsoft (MSN), and their potential utility in creating custom online interactive health maps. Using the programmable interfaces provided by Google and MSN, we created three interactive demonstrator maps of England's Strategic Health Authorities. These can be browsed online at – Google Maps API (Application Programming Interface) version, – Google Earth KML (Keyhole Markup Language) version, and – MSN Virtual Earth Map Control version. Google and MSN's worldwide distribution of "free" geospatial tools, imagery, and maps is to be commended as a significant step towards the ultimate "wikification" of maps and GIS. A discussion is provided of these emerging online mapping trends, their expected future implications and development directions, and associated individual privacy, national security and copyrights issues. Although ESRI have announced their planned response to Google (and MSN), it remains to be seen how their envisaged plans will materialize and compare to the offerings from Google and MSN, and also how Google and MSN mapping tools will further evolve in the near future. PMID:16176577

  2. Validation results of specifications for motion control interoperability

    NASA Astrophysics Data System (ADS)

    Szabo, Sandor; Proctor, Frederick M.

    1997-01-01

    The National Institute of Standards and Technology (NIST) is participating in the Department of Energy Technologies Enabling Agile Manufacturing (TEAM) program to establish interface standards for machine tool, robot, and coordinate measuring machine controllers. At NIST, the focus is to validate potential application programming interfaces (APIs) that make it possible to exchange machine controller components with a minimal impact on the rest of the system. This validation is taking place in the enhanced machine controller (EMC) consortium and is in cooperation with users and vendors of motion control equipment. An area of interest is motion control, including closed-loop control of individual axes and coordinated path planning. Initial tests of the motion control APIs are complete. The APIs were implemented on two commercial motion control boards that run on two different machine tools. The results for a baseline set of APIs look promising, but several issues were raised. These include resolving differing approaches in how motions are programmed and defining a standard measurement of performance for motion control. This paper starts with a summary of the process used in developing a set of specifications for motion control interoperability. Next, the EMC architecture and its classification of motion control APIs into two classes, Servo Control and Trajectory Planning, are reviewed. Selected APIs are presented to explain the basic functionality and some of the major issues involved in porting the APIs to other motion controllers. The paper concludes with a summary of the main issues and ways to continue the standards process.

  3. Using Cloud-based Storage Technologies for Earth Science Data

    NASA Astrophysics Data System (ADS)

    Michaelis, A.; Readey, J.; Votava, P.

    2016-12-01

    Cloud based infrastructure may offer several key benefits of scalability, built in redundancy and reduced total cost of ownership as compared with a traditional data center approach. However, most of the tools and software systems developed for NASA data repositories were not developed with a cloud based infrastructure in mind and do not fully take advantage of commonly available cloud-based technologies. Object storage services are provided through all the leading public (Amazon Web Service, Microsoft Azure, Google Cloud, etc.) and private (Open Stack) clouds, and may provide a more cost-effective means of storing large data collections online. We describe a system that utilizes object storage rather than traditional file system based storage to vend earth science data. The system described is not only cost effective, but shows superior performance for running many different analytics tasks in the cloud. To enable compatibility with existing tools and applications, we outline client libraries that are API compatible with existing libraries for HDF5 and NetCDF4. Performance of the system is demonstrated using clouds services running on Amazon Web Services.

  4. Synergistic effects of non-Apis bees and honey bees for pollination services

    PubMed Central

    Brittain, Claire; Williams, Neal; Kremen, Claire; Klein, Alexandra-Maria

    2013-01-01

    In diverse pollinator communities, interspecific interactions may modify the behaviour and increase the pollination effectiveness of individual species. Because agricultural production reliant on pollination is growing, improving pollination effectiveness could increase crop yield without any increase in agricultural intensity or area. In California almond, a crop highly dependent on honey bee pollination, we explored the foraging behaviour and pollination effectiveness of honey bees in orchards with simple (honey bee only) and diverse (non-Apis bees present) bee communities. In orchards with non-Apis bees, the foraging behaviour of honey bees changed and the pollination effectiveness of a single honey bee visit was greater than in orchards where non-Apis bees were absent. This change translated to a greater proportion of fruit set in these orchards. Our field experiments show that increased pollinator diversity can synergistically increase pollination service, through species interactions that alter the behaviour and resulting functional quality of a dominant pollinator species. These results of functional synergy between species were supported by an additional controlled cage experiment with Osmia lignaria and Apis mellifera. Our findings highlight a largely unexplored facilitative component of the benefit of biodiversity to ecosystem services, and represent a way to improve pollinator-dependent crop yields in a sustainable manner. PMID:23303545

  5. DOE Office of Scientific and Technical Information (OSTI.GOV)

    David Lawrence

    Calibrations and conditions databases can be accessed from within the JANA Event Processing framework through the API defined in its JCalibration base class. The API is designed to support everything from databases, to web services to flat files for the backend. A Web Service backend using the gSOAP toolkit has been implemented which is particularly interesting since it addresses many modern cybersecurity issues including support for SSL. The API allows constants to be retrieved through a single line of C++ code with most of the context, including the transport mechanism, being implied by the run currently being analyzed and themore » environment relieving developers from implementing such details.« less

  6. The caCORE Software Development Kit: streamlining construction of interoperable biomedical information services.

    PubMed

    Phillips, Joshua; Chilukuri, Ram; Fragoso, Gilberto; Warzel, Denise; Covitz, Peter A

    2006-01-06

    Robust, programmatically accessible biomedical information services that syntactically and semantically interoperate with other resources are challenging to construct. Such systems require the adoption of common information models, data representations and terminology standards as well as documented application programming interfaces (APIs). The National Cancer Institute (NCI) developed the cancer common ontologic representation environment (caCORE) to provide the infrastructure necessary to achieve interoperability across the systems it develops or sponsors. The caCORE Software Development Kit (SDK) was designed to provide developers both within and outside the NCI with the tools needed to construct such interoperable software systems. The caCORE SDK requires a Unified Modeling Language (UML) tool to begin the development workflow with the construction of a domain information model in the form of a UML Class Diagram. Models are annotated with concepts and definitions from a description logic terminology source using the Semantic Connector component. The annotated model is registered in the Cancer Data Standards Repository (caDSR) using the UML Loader component. System software is automatically generated using the Codegen component, which produces middleware that runs on an application server. The caCORE SDK was initially tested and validated using a seven-class UML model, and has been used to generate the caCORE production system, which includes models with dozens of classes. The deployed system supports access through object-oriented APIs with consistent syntax for retrieval of any type of data object across all classes in the original UML model. The caCORE SDK is currently being used by several development teams, including by participants in the cancer biomedical informatics grid (caBIG) program, to create compatible data services. caBIG compatibility standards are based upon caCORE resources, and thus the caCORE SDK has emerged as a key enabling technology for caBIG. The caCORE SDK substantially lowers the barrier to implementing systems that are syntactically and semantically interoperable by providing workflow and automation tools that standardize and expedite modeling, development, and deployment. It has gained acceptance among developers in the caBIG program, and is expected to provide a common mechanism for creating data service nodes on the data grid that is under development.

  7. "Dark" Singlet Oxygen and Electron Paramagnetic Resonance Spin Trapping as Convenient Tools to Assess Photolytic Drug Degradation.

    PubMed

    Persich, Peter; Hostyn, Steven; Joie, Céline; Winderickx, Guy; Pikkemaat, Jeroen; Romijn, Edwin P; Maes, Bert U W

    2017-05-01

    Forced degradation studies are an important tool for a systematic assessment of decomposition pathways and identification of reactive sites in active pharmaceutical ingredients (APIs). Two methodologies have been combined in order to provide a deeper understanding of singlet oxygen-related degradation pathways of APIs under light irradiation. First, we report that a "dark" singlet oxygen test enables the investigation of drug reactivity toward singlet oxygen independently of photolytic irradiation processes. Second, the photosensitizing properties of the API producing the singlet oxygen was proven and quantified by spin trapping and electron paramagnetic resonance analysis. A combination of these techniques is an interesting addition to the forced degradation portfolio as it can be used for (1) revealing unexpected degradation pathways of APIs due to singlet oxygen, (2) clarifying photolytic drug-drug interactions in fixed-dose combinations, and (3) synthesizing larger quantities of hardly accessible oxidative drug degradants. Copyright © 2017 American Pharmacists Association®. Published by Elsevier Inc. All rights reserved.

  8. PrismTech Data Distribution Service Java API Evaluation

    NASA Technical Reports Server (NTRS)

    Riggs, Cortney

    2008-01-01

    My internship duties with Launch Control Systems required me to start performance testing of an Object Management Group's (OMG) Data Distribution Service (DDS) specification implementation by PrismTech Limited through the Java programming language application programming interface (API). DDS is a networking middleware for Real-Time Data Distribution. The performance testing involves latency, redundant publishers, extended duration, redundant failover, and read performance. Time constraints allowed only for a data throughput test. I have designed the testing applications to perform all performance tests when time is allowed. Performance evaluation data such as megabits per second and central processing unit (CPU) time consumption were not easily attainable through the Java programming language; they required new methods and classes created in the test applications. Evaluation of this product showed the rate that data can be sent across the network. Performance rates are better on Linux platforms than AIX and Sun platforms. Compared to previous C++ programming language API, the performance evaluation also shows the language differences for the implementation. The Java API of the DDS has a lower throughput performance than the C++ API.

  9. In-line UV spectroscopy for the quantification of low-dose active ingredients during the manufacturing of pharmaceutical semi-solid and liquid formulations.

    PubMed

    Bostijn, N; Hellings, M; Van Der Veen, M; Vervaet, C; De Beer, T

    2018-07-12

    UltraViolet (UV) spectroscopy was evaluated as an innovative Process Analytical Technology (PAT) - tool for the in-line and real-time quantitative determination of low-dosed active pharmaceutical ingredients (APIs) in a semi-solid (gel) and a liquid (suspension) pharmaceutical formulation during their batch production process. The performance of this new PAT-tool (i.e., UV spectroscopy) was compared with an already more established PAT-method based on Raman spectroscopy. In-line UV measurements were carried out with an immersion probe while for the Raman measurements a non-contact PhAT probe was used. For both studied formulations, an in-line API quantification model was developed and validated per spectroscopic technique. The known API concentrations (Y) were correlated with the corresponding in-line collected preprocessed spectra (X) through a Partial Least Squares (PLS) regression. Each developed quantification method was validated by calculating the accuracy profile on the basis of the validation experiments. Furthermore, the measurement uncertainty was determined based on the data generated for the determination of the accuracy profiles. From the accuracy profile of the UV- and Raman-based quantification method for the gel, it was concluded that at the target API concentration of 2% (w/w), 95 out of 100 future routine measurements given by the Raman method will not deviate more than 10% (relative error) from the true API concentration, whereas for the UV method the acceptance limits of 10% were exceeded. For the liquid formulation, the Raman method was not able to quantify the API in the low-dosed suspension (0.09% (w/w) API). In contrast, the in-line UV method was able to adequately quantify the API in the suspension. This study demonstrated that UV spectroscopy can be adopted as a novel in-line PAT-technique for low-dose quantification purposes in pharmaceutical processes. Important is that none of the two spectroscopic techniques was superior to the other for both formulations: the Raman method was more accurate in quantifying the API in the gel (2% (w/w) API), while the UV method performed better for API quantification in the suspension (0.09% (w/w) API). Copyright © 2018 Elsevier B.V. All rights reserved.

  10. Asian Pacific Islander dementia care network: a model of care for underserved communities.

    PubMed

    Kally, Zina; Cherry, Debra L; Howland, Susan; Villarruel, Monica

    2014-01-01

    This study presents the results of the work of the Asian Pacific Islander Dementia Care Network (APIDCN)--a collaborative model of care created to develop community capacity to deliver dementia capable services, build community awareness about Alzheimer's disease and other dementias, and offer direct services to caregivers in the API community in Los Angeles. Through trainings, mentoring, and outreach campaigns, the APIDCN expanded the availability of culturally competent services in the API community. The knowledge that was embedded within partner organizations and in the community at large assures sustainability of the services after the project ended.

  11. WMT: The CSDMS Web Modeling Tool

    NASA Astrophysics Data System (ADS)

    Piper, M.; Hutton, E. W. H.; Overeem, I.; Syvitski, J. P.

    2015-12-01

    The Community Surface Dynamics Modeling System (CSDMS) has a mission to enable model use and development for research in earth surface processes. CSDMS strives to expand the use of quantitative modeling techniques, promotes best practices in coding, and advocates for the use of open-source software. To streamline and standardize access to models, CSDMS has developed the Web Modeling Tool (WMT), a RESTful web application with a client-side graphical interface and a server-side database and API that allows users to build coupled surface dynamics models in a web browser on a personal computer or a mobile device, and run them in a high-performance computing (HPC) environment. With WMT, users can: Design a model from a set of components Edit component parameters Save models to a web-accessible server Share saved models with the community Submit runs to an HPC system Download simulation results The WMT client is an Ajax application written in Java with GWT, which allows developers to employ object-oriented design principles and development tools such as Ant, Eclipse and JUnit. For deployment on the web, the GWT compiler translates Java code to optimized and obfuscated JavaScript. The WMT client is supported on Firefox, Chrome, Safari, and Internet Explorer. The WMT server, written in Python and SQLite, is a layered system, with each layer exposing a web service API: wmt-db: database of component, model, and simulation metadata and output wmt-api: configure and connect components wmt-exe: launch simulations on remote execution servers The database server provides, as JSON-encoded messages, the metadata for users to couple model components, including descriptions of component exchange items, uses and provides ports, and input parameters. Execution servers are network-accessible computational resources, ranging from HPC systems to desktop computers, containing the CSDMS software stack for running a simulation. Once a simulation completes, its output, in NetCDF, is packaged and uploaded to a data server where it is stored and from which a user can download it as a single compressed archive file.

  12. Home Energy Scoring Tools (website) and Application Programming Interfaces, APIs (aka HEScore)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mills, Evan; Bourassa, Norm; Rainer, Leo

    A web-based residential energy rating tool with APIs that runs the LBNL website: Provides customized estimates of residential energy use and energy bills based on building description information provided by the user. Energy use is estimated using engineering models developed at LBNL. Space heating and cooling use is based on the DOE-2. 1E building simulation model. Other end-users (water heating, appliances, lighting, and misc. equipment) are based on engineering models developed by LBNL.

  13. Home Energy Scoring Tools (website) and Application Programming Interfaces, APIs (aka HEScore)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mills, Evan; Bourassa, Norm; Rainer, Leo

    2016-04-22

    A web-based residential energy rating tool with APIs that runs the LBNL website: Provides customized estimates of residential energy use and energy bills based on building description information provided by the user. Energy use is estimated using engineering models developed at LBNL. Space heating and cooling use is based on the DOE-2. 1E building simulation model. Other end-users (water heating, appliances, lighting, and misc. equipment) are based on engineering models developed by LBNL.

  14. The geospatial data quality REST API for primary biodiversity data

    PubMed Central

    Otegui, Javier; Guralnick, Robert P.

    2016-01-01

    Summary: We present a REST web service to assess the geospatial quality of primary biodiversity data. It enables access to basic and advanced functions to detect completeness and consistency issues as well as general errors in the provided record or set of records. The API uses JSON for data interchange and efficient parallelization techniques for fast assessments of large datasets. Availability and implementation: The Geospatial Data Quality API is part of the VertNet set of APIs. It can be accessed at http://api-geospatial.vertnet-portal.appspot.com/geospatial and is already implemented in the VertNet data portal for quality reporting. Source code is freely available under GPL license from http://www.github.com/vertnet/api-geospatial. Contact: javier.otegui@gmail.com or rguralnick@flmnh.ufl.edu Supplementary information: Supplementary data are available at Bioinformatics online. PMID:26833340

  15. The geospatial data quality REST API for primary biodiversity data.

    PubMed

    Otegui, Javier; Guralnick, Robert P

    2016-06-01

    We present a REST web service to assess the geospatial quality of primary biodiversity data. It enables access to basic and advanced functions to detect completeness and consistency issues as well as general errors in the provided record or set of records. The API uses JSON for data interchange and efficient parallelization techniques for fast assessments of large datasets. The Geospatial Data Quality API is part of the VertNet set of APIs. It can be accessed at http://api-geospatial.vertnet-portal.appspot.com/geospatial and is already implemented in the VertNet data portal for quality reporting. Source code is freely available under GPL license from http://www.github.com/vertnet/api-geospatial javier.otegui@gmail.com or rguralnick@flmnh.ufl.edu Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.

  16. 47 CFR 61.43 - Annual price cap filings required.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... tariff year, that make appropriate adjustments to their PCI, API, and SBI values pursuant to §§ 61.45 through 61.47, and that incorporate new services into the PCI, API, or SBI calculations pursuant to §§ 61...

  17. 47 CFR 61.43 - Annual price cap filings required.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... tariff year, that make appropriate adjustments to their PCI, API, and SBI values pursuant to §§ 61.45 through 61.47, and that incorporate new services into the PCI, API, or SBI calculations pursuant to §§ 61...

  18. 47 CFR 61.43 - Annual price cap filings required.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... tariff year, that make appropriate adjustments to their PCI, API, and SBI values pursuant to §§ 61.45 through 61.47, and that incorporate new services into the PCI, API, or SBI calculations pursuant to §§ 61...

  19. 47 CFR 61.43 - Annual price cap filings required.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... tariff year, that make appropriate adjustments to their PCI, API, and SBI values pursuant to §§ 61.45 through 61.47, and that incorporate new services into the PCI, API, or SBI calculations pursuant to §§ 61...

  20. Programmatic access to logical models in the Cell Collective modeling environment via a REST API.

    PubMed

    Kowal, Bryan M; Schreier, Travis R; Dauer, Joseph T; Helikar, Tomáš

    2016-01-01

    Cell Collective (www.cellcollective.org) is a web-based interactive environment for constructing, simulating and analyzing logical models of biological systems. Herein, we present a Web service to access models, annotations, and simulation data in the Cell Collective platform through the Representational State Transfer (REST) Application Programming Interface (API). The REST API provides a convenient method for obtaining Cell Collective data through almost any programming language. To ensure easy processing of the retrieved data, the request output from the API is available in a standard JSON format. The Cell Collective REST API is freely available at http://thecellcollective.org/tccapi. All public models in Cell Collective are available through the REST API. For users interested in creating and accessing their own models through the REST API first need to create an account in Cell Collective (http://thecellcollective.org). thelikar2@unl.edu. Technical user documentation: https://goo.gl/U52GWo. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  1. The AppScale Cloud Platform

    PubMed Central

    Krintz, Chandra

    2013-01-01

    AppScale is an open source distributed software system that implements a cloud platform as a service (PaaS). AppScale makes cloud applications easy to deploy and scale over disparate cloud fabrics, implementing a set of APIs and architecture that also makes apps portable across the services they employ. AppScale is API-compatible with Google App Engine (GAE) and thus executes GAE applications on-premise or over other cloud infrastructures, without modification. PMID:23828721

  2. The IRMIS object model and services API.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Saunders, C.; Dohan, D. A.; Arnold, N. D.

    2005-01-01

    The relational model developed for the Integrated Relational Model of Installed Systems (IRMIS) toolkit has been successfully used to capture the Advanced Photon Source (APS) control system software (EPICS process variables and their definitions). The relational tables are populated by a crawler script that parses each Input/Output Controller (IOC) start-up file when an IOC reboot is detected. User interaction is provided by a Java Swing application that acts as a desktop for viewing the process variable information. Mapping between the display objects and the relational tables was carried out with the Hibernate Object Relational Modeling (ORM) framework. Work is wellmore » underway at the APS to extend the relational modeling to include control system hardware. For this work, due in part to the complex user interaction required, the primary application development environment has shifted from the relational database view to the object oriented (Java) perspective. With this approach, the business logic is executed in Java rather than in SQL stored procedures. This paper describes the object model used to represent control system software, hardware, and interconnects in IRMIS. We also describe the services API used to encapsulate the required behaviors for creating and maintaining the complex data. In addition to the core schema and object model, many important concepts in IRMIS are captured by the services API. IRMIS is an ambitious collaborative effort for defining and developing a relational database and associated applications to comprehensively document the large and complex EPICS-based control systems of today's accelerators. The documentation effort includes process variables, control system hardware, and interconnections. The approach could also be used to document all components of the accelerator, including mechanical, vacuum, power supplies, etc. One key aspect of IRMIS is that it is a documentation framework, not a design and development tool. We do not generate EPICS control system configurations from IRMIS, and hence do not impose any additional requirements on EPICS developers.« less

  3. Ionic liquids in drug delivery.

    PubMed

    Shamshina, Julia L; Barber, Patrick S; Rogers, Robin D

    2013-10-01

    To overcome potential problems with solid-state APIs, such as polymorphism, solubility and bioavailability, pure liquid salt (ionic liquid) forms of active pharmaceutical ingredients (API-ILs) are considered here as a design strategy. After a critical review of the current literature, the recent development of the API-ILs strategy is presented, with a particular focus on the liquefaction of drugs. A variety of IL tools for control over the liquid salt state of matter are discussed including choice of counterion to produce an IL from a given API; the concept of oligomeric ions that enables liquefaction of solid ILs by changing the stoichiometry or complexity of the ions; formation of 'liquid co-crystals' where hydrogen bonding is the driving force in the liquefaction of a neutral acid-base complex; combining an IL strategy with the prodrug strategy to improve the delivery of solid APIs; using ILs as delivery agents via trapping a drug in a micelle and finally ILs designed with tunable hydrophilic-lipophilic balance that matches the structural requirements needed to solubilize poorly water-soluble APIs. The authors believe that API-IL approaches may save failed lead candidates, extend the patent life of current APIs, lead to new delivery options or even new pharmaceutical action. They encourage the pharmaceutical industry to invest more research into the API-IL platform as it could lead to fast-tracked approval based on similarities to the APIs already approved.

  4. A Highly Scalable Data Service (HSDS) using Cloud-based Storage Technologies for Earth Science Data

    NASA Astrophysics Data System (ADS)

    Michaelis, A.; Readey, J.; Votava, P.; Henderson, J.; Willmore, F.

    2017-12-01

    Cloud based infrastructure may offer several key benefits of scalability, built in redundancy, security mechanisms and reduced total cost of ownership as compared with a traditional data center approach. However, most of the tools and legacy software systems developed for online data repositories within the federal government were not developed with a cloud based infrastructure in mind and do not fully take advantage of commonly available cloud-based technologies. Moreover, services bases on object storage are well established and provided through all the leading cloud service providers (Amazon Web Service, Microsoft Azure, Google Cloud, etc…) of which can often provide unmatched "scale-out" capabilities and data availability to a large and growing consumer base at a price point unachievable from in-house solutions. We describe a system that utilizes object storage rather than traditional file system based storage to vend earth science data. The system described is not only cost effective, but shows a performance advantage for running many different analytics tasks in the cloud. To enable compatibility with existing tools and applications, we outline client libraries that are API compatible with existing libraries for HDF5 and NetCDF4. Performance of the system is demonstrated using clouds services running on Amazon Web Services.

  5. deepTools2: a next generation web server for deep-sequencing data analysis.

    PubMed

    Ramírez, Fidel; Ryan, Devon P; Grüning, Björn; Bhardwaj, Vivek; Kilpert, Fabian; Richter, Andreas S; Heyne, Steffen; Dündar, Friederike; Manke, Thomas

    2016-07-08

    We present an update to our Galaxy-based web server for processing and visualizing deeply sequenced data. Its core tool set, deepTools, allows users to perform complete bioinformatic workflows ranging from quality controls and normalizations of aligned reads to integrative analyses, including clustering and visualization approaches. Since we first described our deepTools Galaxy server in 2014, we have implemented new solutions for many requests from the community and our users. Here, we introduce significant enhancements and new tools to further improve data visualization and interpretation. deepTools continue to be open to all users and freely available as a web service at deeptools.ie-freiburg.mpg.de The new deepTools2 suite can be easily deployed within any Galaxy framework via the toolshed repository, and we also provide source code for command line usage under Linux and Mac OS X. A public and documented API for access to deepTools functionality is also available. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  6. Simulation Data Management - Requirements and Design Specification

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Clay, Robert L.; Friedman-Hill, Ernest J.; Gibson, Marcus J.

    Simulation Data Management (SDM), the ability to securely organize, archive, and share analysis models and the artifacts used to create them, is a fundamental requirement for modern engineering analysis based on computational simulation. We have worked separately to provide secure, network SDM services to engineers and scientists at our respective laboratories for over a decade. We propose to leverage our experience and lessons learned to help develop and deploy a next-generation SDM service as part of a multi-laboratory team. This service will be portable across multiple sites and platforms, and will be accessible via a range of command-line tools andmore » well-documented APIs. In this document, we’ll review our high-level and low-level requirements for such a system, review one existing system, and briefly discuss our proposed implementation.« less

  7. Mercury: Reusable software application for Metadata Management, Data Discovery and Access

    NASA Astrophysics Data System (ADS)

    Devarakonda, Ranjeet; Palanisamy, Giri; Green, James; Wilson, Bruce E.

    2009-12-01

    Mercury is a federated metadata harvesting, data discovery and access tool based on both open source packages and custom developed software. It was originally developed for NASA, and the Mercury development consortium now includes funding from NASA, USGS, and DOE. Mercury is itself a reusable toolset for metadata, with current use in 12 different projects. Mercury also supports the reuse of metadata by enabling searching across a range of metadata specification and standards including XML, Z39.50, FGDC, Dublin-Core, Darwin-Core, EML, and ISO-19115. Mercury provides a single portal to information contained in distributed data management systems. It collects metadata and key data from contributing project servers distributed around the world and builds a centralized index. The Mercury search interfaces then allow the users to perform simple, fielded, spatial and temporal searches across these metadata sources. One of the major goals of the recent redesign of Mercury was to improve the software reusability across the projects which currently fund the continuing development of Mercury. These projects span a range of land, atmosphere, and ocean ecological communities and have a number of common needs for metadata searches, but they also have a number of needs specific to one or a few projects To balance these common and project-specific needs, Mercury’s architecture includes three major reusable components; a harvester engine, an indexing system and a user interface component. The harvester engine is responsible for harvesting metadata records from various distributed servers around the USA and around the world. The harvester software was packaged in such a way that all the Mercury projects will use the same harvester scripts but each project will be driven by a set of configuration files. The harvested files are then passed to the Indexing system, where each of the fields in these structured metadata records are indexed properly, so that the query engine can perform simple, keyword, spatial and temporal searches across these metadata sources. The search user interface software has two API categories; a common core API which is used by all the Mercury user interfaces for querying the index and a customized API for project specific user interfaces. For our work in producing a reusable, portable, robust, feature-rich application, Mercury received a 2008 NASA Earth Science Data Systems Software Reuse Working Group Peer-Recognition Software Reuse Award. The new Mercury system is based on a Service Oriented Architecture and effectively reuses components for various services such as Thesaurus Service, Gazetteer Web Service and UDDI Directory Services. The software also provides various search services including: RSS, Geo-RSS, OpenSearch, Web Services and Portlets, integrated shopping cart to order datasets from various data centers (ORNL DAAC, NSIDC) and integrated visualization tools. Other features include: Filtering and dynamic sorting of search results, book-markable search results, save, retrieve, and modify search criteria.

  8. Vocabulary services to support scientific data interoperability

    NASA Astrophysics Data System (ADS)

    Cox, Simon; Mills, Katie; Tan, Florence

    2013-04-01

    Shared vocabularies are a core element in interoperable systems. Vocabularies need to be available at run-time, and where the vocabularies are shared by a distributed community this implies the use of web technology to provide vocabulary services. Given the ubiquity of vocabularies or classifiers in systems, vocabulary services are effectively the base of the interoperability stack. In contemporary knowledge organization systems, a vocabulary item is considered a concept, with the "terms" denoting it appearing as labels. The Simple Knowledge Organization System (SKOS) formalizes this as an RDF Schema (RDFS) application, with a bridge to formal logic in Web Ontology Language (OWL). For maximum utility, a vocabulary should be made available through the following interfaces: * the vocabulary as a whole - at an ontology URI corresponding to a vocabulary document * each item in the vocabulary - at the item URI * summaries, subsets, and resources derived by transformation * through the standard RDF web API - i.e. a SPARQL endpoint * through a query form for human users. However, the vocabulary data model may be leveraged directly in a standard vocabulary API that uses the semantics provided by SKOS. SISSvoc3 [1] accomplishes this as a standard set of URI templates for a vocabulary. Any URI comforming to the template selects a vocabulary subset based on the SKOS properties, including labels (skos:prefLabel, skos:altLabel, rdfs:label) and a subset of the semantic relations (skos:broader, skos:narrower, etc). SISSvoc3 thus provides a RESTFul SKOS API to query a vocabulary, but hiding the complexity of SPARQL. It has been implemented using the Linked Data API (LDA) [2], which connects to a SPARQL endpoint. By using LDA, we also get content-negotiation, alternative views, paging, metadata and other functionality provided in a standard way. A number of vocabularies have been formalized in SKOS and deployed by CSIRO, the Australian Bureau of Meteorology (BOM) and their collaborators using SISSvoc3, including: * geologic timescale (multiple versions) * soils classification * definitions from OGC standards * geosciml vocabularies * mining commodities * hyperspectral scalars Several other agencies in Australia have adopted SISSvoc3 for their vocabularies. SISSvoc3 differs from other SKOS-based vocabulary-access APIs such as GEMET [3] and NVS [4] in that (a) the service is decoupled from the content store, (b) the service URI is independent of the content URIs This means that a SISSvoc3 interface can be deployed over any SKOS vocabulary which is available at a SPARQL endpoint. As an example, a SISSvoc3 query and presentation interface has been deployed over the NERC vocabulary service hosted by the BODC, providing a search interface which is not available natively. We use vocabulary services to populate menus in user interfaces, to support data validation, and to configure data conversion routines. Related services built on LDA have also been used as a generic registry interface, and extended for serving gazetteer information. ACKNOWLEDGEMENTS The CSIRO SISSvoc3 implementation is built using the Epimorphics ELDA platform http://code.google.com/p/elda/. We thank Jacqui Githaiga and Terry Rankine for their contributions to SISSvoc design and implementation. REFERENCES 1. SISSvoc3 Specification https://www.seegrid.csiro.au/wiki/Siss/SISSvoc30Specification 2. Linked Data API http://code.google.com/p/linked-data-api/wiki/Specification 3. GEMET https://svn.eionet.europa.eu/projects/Zope/wiki/GEMETWebServiceAPI 4. NVS 2.0 http://vocab.nerc.ac.uk/

  9. Scientific Use Cases for the Virtual Atomic and Molecular Data Center

    NASA Astrophysics Data System (ADS)

    Dubernet, M. L.; Aboudarham, J.; Ba, Y. A.; Boiziot, M.; Bottinelli, S.; Caux, E.; Endres, C.; Glorian, J. M.; Henry, F.; Lamy, L.; Le Sidaner, P.; Møller, T.; Moreau, N.; Rénié, C.; Roueff, E.; Schilke, P.; Vastel, C.; Zwoelf, C. M.

    2014-12-01

    VAMDC Consortium is a worldwide consortium which federates interoperable Atomic and Molecular databases through an e-science infrastructure. The contained data are of the highest scientific quality and are crucial for many applications: astrophysics, atmospheric physics, fusion, plasma and lighting technologies, health, etc. In this paper we present astrophysical scientific use cases in relation to the use of the VAMDC e-infrastructure. Those will cover very different applications such as: (i) modeling the spectra of interstellar objects using the myXCLASS software tool implemented in the Common Astronomy Software Applications package (CASA) or using the CASSIS software tool, in its stand-alone version or implemented in the Herschel Interactive Processing Environment (HIPE); (ii) the use of Virtual Observatory tools accessing VAMDC databases; (iii) the access of VAMDC from the Paris solar BASS2000 portal; (iv) the combination of tools and database from the APIS service (Auroral Planetary Imaging and Spectroscopy); (v) combination of heterogeneous data for the application to the interstellar medium from the SPECTCOL tool.

  10. The Earth Observatory Natural Event Tracker (EONET): An API for Matching Natural Events to GIBS Imagery

    NASA Astrophysics Data System (ADS)

    Ward, K.

    2015-12-01

    Hidden within the terabytes of imagery in NASA's Global Imagery Browse Services (GIBS) collection are hundreds of daily natural events. Some events are newsworthy, devastating, and visibly obvious at a global scale, others are merely regional curiosities. Regardless of the scope and significance of any one event, it is likely that multiple GIBS layers can be viewed to provide a multispectral, dataset-based view of the event. To facilitate linking between the discrete event and the representative dataset imagery, NASA's Earth Observatory Group has developed a prototype application programming interface (API): the Earth Observatory Natural Event Tracker (EONET). EONET supports an API model that allows users to retrieve event-specific metadata--date/time, location, and type (wildfire, storm, etc.)--and web service layer-specific metadata which can be used to link to event-relevant dataset imagery in GIBS. GIBS' ability to ingest many near real time datasets, combined with its growing archive of past imagery, means that API users will be able to develop client applications that not only show ongoing events but can also look at imagery from before and after. In our poster, we will present the API and show examples of its use.

  11. Software for Remote Monitoring of Space-Station Payloads

    NASA Technical Reports Server (NTRS)

    Schneider, Michelle; Lippincott, Jeff; Chubb, Steve; Whitaker, Jimmy; Gillis, Robert; Sellers, Donna; Sims, Chris; Rice, James

    2003-01-01

    Telescience Resource Kit (TReK) is a suite of application programs that enable geographically dispersed users to monitor scientific payloads aboard the International Space Station (ISS). TReK provides local ground support services that can simultaneously receive, process, record, playback, and display data from multiple sources. TReK also provides interfaces to use the remote services provided by the Payload Operations Integration Center which manages all ISS payloads. An application programming interface (API) allows for payload users to gain access to all data processed by TReK and allows payload-specific tools and programs to be built or integrated with TReK. Used in conjunction with other ISS-provided tools, TReK provides the ability to integrate payloads with the operational ground system early in the lifecycle. This reduces the potential for operational problems and provides "cradle-to-grave" end-to-end operations. TReK contains user guides and self-paced tutorials along with training applications to allow the user to become familiar with the system.

  12. IgE-Api m 4 Is Useful for Identifying a Particular Phenotype of Bee Venom Allergy.

    PubMed

    Ruiz, B; Serrano, P; Moreno, C

    Different clinical behaviors have been identified in patients allergic to bee venom. Compound-resolved diagnosis could be an appropriate tool for investigating these differences. The aims of this study were to analyze whether specific IgE to Api m 4 (sIgE-Api m 4) can identify a particular kind of bee venom allergy and to describe response to bee venom immunotherapy (bVIT). Prospective study of 31 patients allergic to bee venom who were assigned to phenotype group A (sIgE-Api m 4 <0.98 kU/L), treated with native aqueous (NA) extract, or phenotype group B (sIgE-Api m 4 ≥0.98 kU/L), treated with purified aqueous (PA) extract. Sex, age, cardiovascular risk, severity of preceding sting reaction, exposure to beekeeping, and immunological data (intradermal test, sIgE/sIgG4-Apis-nApi m 1, and sIgE-rApi m 2-Api m 4 were analyzed. Systemic reactions (SRs) during bVIT build-up were analyzed. Immunological and sting challenge outcomes were evaluated in each group after 1 and 2 years of bVIT. Phenotype B patients had more severe reactions (P=.049) and higher skin sensitivity (P=.011), baseline sIgE-Apis (P=.0004), sIgE-nApi m 1 (P=.0004), and sIgG4-Apis (P=.027) than phenotype A patients. Furthermore, 41% of patients in group B experienced SRs during the build-up phase with NA; the sting challenge success rate in this group was 82%. There were no significant reductions in serial intradermal test results, but an intense reduction in sIgE-nApi m 1 (P=.013) and sIgE-Api m 4 (P=.004) was observed after the first year of bVIT. Use of IgE-Api m 4 as the only discrimination criterion demonstrated differences in bee venom allergy. Further investigation with larger populations is necessary.

  13. Automated Formal Testing of C API Using T2C Framework

    NASA Astrophysics Data System (ADS)

    Khoroshilov, Alexey V.; Rubanov, Vladimir V.; Shatokhin, Eugene A.

    A problem of automated test development for checking basic functionality of program interfaces (API) is discussed. Different technologies and corresponding tools are surveyed. And T2C technology developed in ISPRAS is presented. The technology and associated tools facilitate development of "medium quality" (and "medium cost") tests. An important feature of T2C technology is that it enforces that each check in a developed test is explicitly linked to the corresponding place in the standard. T2C tools provide convenient means to create such linkage. The results of using T2C are considered by example of a project for testing interfaces of Linux system libraries defined by the LSB standard.

  14. The CMS dataset bookkeeping service

    NASA Astrophysics Data System (ADS)

    Afaq, A.; Dolgert, A.; Guo, Y.; Jones, C.; Kosyakov, S.; Kuznetsov, V.; Lueking, L.; Riley, D.; Sekhri, V.

    2008-07-01

    The CMS Dataset Bookkeeping Service (DBS) has been developed to catalog all CMS event data from Monte Carlo and Detector sources. It provides the ability to identify MC or trigger source, track data provenance, construct datasets for analysis, and discover interesting data. CMS requires processing and analysis activities at various service levels and the DBS system provides support for localized processing or private analysis, as well as global access for CMS users at large. Catalog entries can be moved among the various service levels with a simple set of migration tools, thus forming a loose federation of databases. DBS is available to CMS users via a Python API, Command Line, and a Discovery web page interfaces. The system is built as a multi-tier web application with Java servlets running under Tomcat, with connections via JDBC to Oracle or MySQL database backends. Clients connect to the service through HTTP or HTTPS with authentication provided by GRID certificates and authorization through VOMS. DBS is an integral part of the overall CMS Data Management and Workflow Management systems.

  15. Web scraping technologies in an API world.

    PubMed

    Glez-Peña, Daniel; Lourenço, Anália; López-Fernández, Hugo; Reboiro-Jato, Miguel; Fdez-Riverola, Florentino

    2014-09-01

    Web services are the de facto standard in biomedical data integration. However, there are data integration scenarios that cannot be fully covered by Web services. A number of Web databases and tools do not support Web services, and existing Web services do not cover for all possible user data demands. As a consequence, Web data scraping, one of the oldest techniques for extracting Web contents, is still in position to offer a valid and valuable service to a wide range of bioinformatics applications, ranging from simple extraction robots to online meta-servers. This article reviews existing scraping frameworks and tools, identifying their strengths and limitations in terms of extraction capabilities. The main focus is set on showing how straightforward it is today to set up a data scraping pipeline, with minimal programming effort, and answer a number of practical needs. For exemplification purposes, we introduce a biomedical data extraction scenario where the desired data sources, well-known in clinical microbiology and similar domains, do not offer programmatic interfaces yet. Moreover, we describe the operation of WhichGenes and PathJam, two bioinformatics meta-servers that use scraping as means to cope with gene set enrichment analysis. © The Author 2013. Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.

  16. The synergistic effects of almond protection fungicides on honey bee (Apis mellifera) forager survival

    USDA-ARS?s Scientific Manuscript database

    The honey bee (Apis mellifera) contributes approximately $17 billion annually in pollination services performed for major agricultural crops in the United States including almond, which is completely dependent on honey bee pollination for nut set. Almond growers face challenges to crop productivity ...

  17. BamTools: a C++ API and toolkit for analyzing and managing BAM files.

    PubMed

    Barnett, Derek W; Garrison, Erik K; Quinlan, Aaron R; Strömberg, Michael P; Marth, Gabor T

    2011-06-15

    Analysis of genomic sequencing data requires efficient, easy-to-use access to alignment results and flexible data management tools (e.g. filtering, merging, sorting, etc.). However, the enormous amount of data produced by current sequencing technologies is typically stored in compressed, binary formats that are not easily handled by the text-based parsers commonly used in bioinformatics research. We introduce a software suite for programmers and end users that facilitates research analysis and data management using BAM files. BamTools provides both the first C++ API publicly available for BAM file support as well as a command-line toolkit. BamTools was written in C++, and is supported on Linux, Mac OSX and MS Windows. Source code and documentation are freely available at http://github.org/pezmaster31/bamtools.

  18. Leveraging Open Standards and Technologies to Search and Display Planetary Image Data

    NASA Astrophysics Data System (ADS)

    Rose, M.; Schauer, C.; Quinol, M.; Trimble, J.

    2011-12-01

    Mars and the Moon have both been visited by multiple NASA spacecraft. A large number of images and other data have been gathered by the spacecraft and are publicly available in NASA's Planetary Data System. Through a collaboration with Google, Inc., the User Centered Technologies group at NASA Ames Resarch Center has developed at tool for searching and browsing among images from multiple Mars and Moon missions. Development of this tool was facilitated by the use of several open technologies and standards. First, an open-source full-text search engine is used to search both place names on the target and to find images matching a geographic region. Second, the published API of the Google Earth browser plugin is used to geolocate the images on a virtual globe and allow the user to navigate on the globe to see related images. The structure of the application also employs standard protocols and services. The back-end is exposed as RESTful APIs, which could be reused by other client systems in the future. Further, the communication between the front- and back-end portions of the system utilizes open data standards including XML and KML (Keyhole Markup Language) for representation of textual and geographic data. The creation of the search index was facilitated by reuse of existing, publicly available metadata, including the Gazetteer of Planetary Nomenclature from the USGS, available in KML format. And the image metadata was reused from standards-compliant archives in the Planetary Data System. The system also supports collaboration with other tools by allowing export of search results in KML, and the ability to display those results in the Google Earth desktop application. We will demonstrate the search and visualization capabilities of the system, with emphasis on how the system facilitates reuse of data and services through the adoption of open standards.

  19. The caCORE Software Development Kit: Streamlining construction of interoperable biomedical information services

    PubMed Central

    Phillips, Joshua; Chilukuri, Ram; Fragoso, Gilberto; Warzel, Denise; Covitz, Peter A

    2006-01-01

    Background Robust, programmatically accessible biomedical information services that syntactically and semantically interoperate with other resources are challenging to construct. Such systems require the adoption of common information models, data representations and terminology standards as well as documented application programming interfaces (APIs). The National Cancer Institute (NCI) developed the cancer common ontologic representation environment (caCORE) to provide the infrastructure necessary to achieve interoperability across the systems it develops or sponsors. The caCORE Software Development Kit (SDK) was designed to provide developers both within and outside the NCI with the tools needed to construct such interoperable software systems. Results The caCORE SDK requires a Unified Modeling Language (UML) tool to begin the development workflow with the construction of a domain information model in the form of a UML Class Diagram. Models are annotated with concepts and definitions from a description logic terminology source using the Semantic Connector component. The annotated model is registered in the Cancer Data Standards Repository (caDSR) using the UML Loader component. System software is automatically generated using the Codegen component, which produces middleware that runs on an application server. The caCORE SDK was initially tested and validated using a seven-class UML model, and has been used to generate the caCORE production system, which includes models with dozens of classes. The deployed system supports access through object-oriented APIs with consistent syntax for retrieval of any type of data object across all classes in the original UML model. The caCORE SDK is currently being used by several development teams, including by participants in the cancer biomedical informatics grid (caBIG) program, to create compatible data services. caBIG compatibility standards are based upon caCORE resources, and thus the caCORE SDK has emerged as a key enabling technology for caBIG. Conclusion The caCORE SDK substantially lowers the barrier to implementing systems that are syntactically and semantically interoperable by providing workflow and automation tools that standardize and expedite modeling, development, and deployment. It has gained acceptance among developers in the caBIG program, and is expected to provide a common mechanism for creating data service nodes on the data grid that is under development. PMID:16398930

  20. Home - Energy Innovation Portal

    Science.gov Websites

    tree sapling with a single leaf. Browse 1,225 Technology Marketing Summaries Graphic of a small tree . Learn about 17 Success Stories More Features API API Get technology info via a web service. EERE officials present the Clean Energy Challenge award to a young entrepreneur. EERE Technology-to-Market Visit

  1. Development of an in-line Raman spectroscopic method for continuous API quantification during twin-screw wet granulation.

    PubMed

    Harting, Julia; Kleinebudde, Peter

    2018-04-01

    Raman spectroscopy was evaluated as a process analytical technology (PAT) tool for continuous API quantification during twin-screw wet granulation. Therefore, a Raman probe was implemented in front of the granulator barrel. This setup enabled the collection of Raman spectra upon a constant granule flow. To develop an in-line PLS calibration model, eight binary mixtures of the API and lactose monohydrate with API contents between 5 and 50% were pre-blended and granulated in a twin-screw granulator with a screw speed of 150 rpm and a powder feed rate of 40 g/min. Water was used as a granulation liquid with different liquid to solid ratios depending on the API content. Ibuprofen and diclofenac sodium were chosen as model drugs and separated PLS models were built for each API. The predictive performance of the developed PLS models was determined by granulating and monitoring new test samples containing different API concentrations. This evaluation showed that the models were able to predict the API concentration with an RMSEP of 0.59% for ibuprofen and 1.5% for diclofenac sodium. In a second part, the developed in-line Raman spectroscopic method was used to determine the API concentration during a split feeding process. Therefore, the API and lactose monohydrate were added by two independently adjustable feeders into the twin-screw granulator barrel. The in-line spectroscopy analysis which was verified by UV-analysis indicated that the mixing ability of the twin-screw granulator was good for the used settings and all adjusted API concentrations. Copyright © 2018 Elsevier B.V. All rights reserved.

  2. STEREO In-situ Data Analysis

    NASA Astrophysics Data System (ADS)

    Schroeder, P. C.; Luhmann, J. G.; Davis, A. J.; Russell, C. T.

    2006-12-01

    STEREO's IMPACT (In-situ Measurements of Particles and CME Transients) investigation provides the first opportunity for long duration, detailed observations of 1 AU magnetic field structures, plasma and suprathermal electrons, and energetic particles at points bracketing Earth's heliospheric location. The PLASTIC instrument takes plasma ion composition measurements completing STEREO's comprehensive in-situ perspective. Stereoscopic/3D information from the STEREO SECCHI imagers and SWAVES radio experiment make it possible to use both multipoint and quadrature studies to connect interplanetary Coronal Mass Ejections (ICME) and solar wind structures to CMEs and coronal holes observed at the Sun. The uniqueness of the STEREO mission requires novel data analysis tools and techniques to take advantage of the mission's full scientific potential. An interactive browser with the ability to create publication-quality plots has been developed which integrates STEREO's in-situ data with data from a variety of other missions including WIND and ACE. Also, an application program interface (API) is provided allowing users to create custom software that ties directly into STEREO's data set. The API allows for more advanced forms of data mining than currently available through most web-based data services. A variety of data access techniques and the development of cross-spacecraft data analysis tools allow the larger scientific community to combine STEREO's unique in-situ data with those of other missions, particularly the L1 missions, and, therefore, to maximize STEREO's scientific potential in gaining a greater understanding of the heliosphere.

  3. BamTools: a C++ API and toolkit for analyzing and managing BAM files

    PubMed Central

    Barnett, Derek W.; Garrison, Erik K.; Quinlan, Aaron R.; Strömberg, Michael P.; Marth, Gabor T.

    2011-01-01

    Motivation: Analysis of genomic sequencing data requires efficient, easy-to-use access to alignment results and flexible data management tools (e.g. filtering, merging, sorting, etc.). However, the enormous amount of data produced by current sequencing technologies is typically stored in compressed, binary formats that are not easily handled by the text-based parsers commonly used in bioinformatics research. Results: We introduce a software suite for programmers and end users that facilitates research analysis and data management using BAM files. BamTools provides both the first C++ API publicly available for BAM file support as well as a command-line toolkit. Availability: BamTools was written in C++, and is supported on Linux, Mac OSX and MS Windows. Source code and documentation are freely available at http://github.org/pezmaster31/bamtools. Contact: barnetde@bc.edu PMID:21493652

  4. A Knowledge Portal and Collaboration Environment for the Earth Sciences

    NASA Astrophysics Data System (ADS)

    D'Agnese, F. A.

    2008-12-01

    Earth Knowledge is developing a web-based 'Knowledge Portal and Collaboration Environment' that will serve as the information-technology-based foundation of a modular Internet-based Earth-Systems Monitoring, Analysis, and Management Tool. This 'Knowledge Portal' is essentially a 'mash- up' of web-based and client-based tools and services that support on-line collaboration, community discussion, and broad public dissemination of earth and environmental science information in a wide-area distributed network. In contrast to specialized knowledge-management or geographic-information systems developed for long- term and incremental scientific analysis, this system will exploit familiar software tools using industry standard protocols, formats, and APIs to discover, process, fuse, and visualize existing environmental datasets using Google Earth and Google Maps. An early form of these tools and services is being used by Earth Knowledge to facilitate the investigations and conversations of scientists, resource managers, and citizen-stakeholders addressing water resource sustainability issues in the Great Basin region of the desert southwestern United States. These ongoing projects will serve as use cases for the further development of this information-technology infrastructure. This 'Knowledge Portal' will accelerate the deployment of Earth- system data and information into an operational knowledge management system that may be used by decision-makers concerned with stewardship of water resources in the American Desert Southwest.

  5. SAMuS: Service-Oriented Architecture for Multisensor Surveillance in Smart Homes

    PubMed Central

    Van de Walle, Rik

    2014-01-01

    The design of a service-oriented architecture for multisensor surveillance in smart homes is presented as an integrated solution enabling automatic deployment, dynamic selection, and composition of sensors. Sensors are implemented as Web-connected devices, with a uniform Web API. RESTdesc is used to describe the sensors and a novel solution is presented to automatically compose Web APIs that can be applied with existing Semantic Web reasoners. We evaluated the solution by building a smart Kinect sensor that is able to dynamically switch between IR and RGB and optimizing person detection by incorporating feedback from pressure sensors, as such demonstrating the collaboration among sensors to enhance detection of complex events. The performance results show that the platform scales for many Web APIs as composition time remains limited to a few hundred milliseconds in almost all cases. PMID:24778579

  6. Globus: Service and Platform for Research Data Lifecycle Management

    NASA Astrophysics Data System (ADS)

    Ananthakrishnan, R.; Foster, I.

    2017-12-01

    Globus offers a range of data management capabilities to the community as hosted services, encompassing data transfer and sharing, user identity and authorization, and data publication. Globus capabilities are accessible via both a web browser and REST APIs. Web access allows researchers to use Globus capabilities through a software-as-a-service model; and the REST APIs address the needs of developers of research services, who can now use Globus as a platform, outsourcing complex user and data management tasks to Globus services. In this presentation, we review Globus capabilities and outline how it is being applied as a platform for scientific services, and highlight work done to link computational analysis flows to the underlying data through an interactive Jupyter notebook environment to promote immediate data usability, reusability of these flows by other researchers, and future analysis extensibility.

  7. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Klise, Katherine A.; Murray, Regan; Bynum, Michael

    Water utilities are vulnerable to a wide variety of human-caused and natural disasters. These disruptive events can result in loss of water service, contaminated water, pipe breaks, and failed equipment. Furthermore, long term changes in water supply and customer demand can have a large impact on the operating conditions of the network. The ability to maintain drinking water service during and following these types of events is critical. Simulation and analysis tools can help water utilities explore how their network will respond to disruptive events and plan effective mitigation strategies. The U.S. Environmental Protection Agency and Sandia National Laboratories aremore » developing new software tools to meet this need. The Water Network Tool for Resilience (WNTR, pronounced winter) is a Python package designed to help water utilities investigate resilience of water distribution systems over a wide range of hazardous scenarios and to evaluate resilience-enhancing actions. The following documentation includes installation instructions and examples, description of software features, and software license. It is assumed that the reader is familiar with the Python Programming Language. References are included for additional background on software components. Online documentation, hosted at http://wntr.readthedocsio/, will be updated as new features are added. The online version includes API documentation and information for developers.« less

  8. EarthServer2 : The Marine Data Service - Web based and Programmatic Access to Ocean Colour Open Data

    NASA Astrophysics Data System (ADS)

    Clements, Oliver; Walker, Peter

    2017-04-01

    The ESA Ocean Colour - Climate Change Initiative (ESA OC-CCI) has produced a long-term high quality global dataset with associated per-pixel uncertainty data. This dataset has now grown to several hundred terabytes (uncompressed) and is freely available to download. However, the sheer size of the dataset can act as a barrier to many users; large network bandwidth, local storage and processing requirements can prevent researchers without the backing of a large organisation from taking advantage of this raw data. The EC H2020 project, EarthServer2, aims to create a federated data service providing access to more than 1 petabyte of earth science data. Within this federation the Marine Data Service already provides an innovative on-line tool-kit for filtering, analysing and visualising OC-CCI data. Data are made available, filtered and processed at source through a standards-based interface, the Open Geospatial Consortium Web Coverage Service and Web Coverage Processing Service. This work was initiated in the EC FP7 EarthServer project where it was found that the unfamiliarity and complexity of these interfaces itself created a barrier to wider uptake. The continuation project, EarthServer2, addresses these issues by providing higher level tools for working with these data. We will present some examples of these tools. Many researchers wish to extract time series data from discrete points of interest. We will present a web based interface, based on NASA/ESA WebWorldWind, for selecting points of interest and plotting time series from a chosen dataset. In addition, a CSV file of locations and times, such as a ship's track, can be uploaded and these points extracted and returned in a CSV file allowing researchers to work with the extract locally, such as a spreadsheet. We will also present a set of Python and JavaScript APIs that have been created to complement and extend the web based GUI. These APIs allow the selection of single points and areas for extraction. The extracted data is returned as structured data (for instance a Python array) which can then be passed directly to local processing code. We will highlight how the libraries can be used by the community and integrated into existing systems, for instance by the use of Jupyter notebooks to share Python code examples which can then be used by other researchers as a basis for their own work.

  9. On-line near infrared spectroscopy as a Process Analytical Technology (PAT) tool to control an industrial seeded API crystallization.

    PubMed

    Schaefer, C; Lecomte, C; Clicq, D; Merschaert, A; Norrant, E; Fotiadu, F

    2013-09-01

    The final step of an active pharmaceutical ingredient (API) manufacturing synthesis process consists of a crystallization during which the API and residual solvent contents have to be quantified precisely in order to reach a predefined seeding point. A feasibility study was conducted to demonstrate the suitability of on-line NIR spectroscopy to control this step in line with new version of the European Medicines Agency (EMA) guideline [1]. A quantitative method was developed at laboratory scale using statistical design of experiments (DOE) and multivariate data analysis such as principal component analysis (PCA) and partial least squares (PLS) regression. NIR models were built to quantify the API in the range of 9-12% (w/w) and to quantify the residual methanol in the range of 0-3% (w/w). To improve the predictive ability of the models, the development procedure encompassed: outliers elimination, optimum model rank definition, spectral range and spectral pre-treatment selection. Conventional criteria such as, number of PLS factors, R(2), root mean square errors of calibration, cross-validation and prediction (RMSEC, RMSECV, RMSEP) enabled the selection of three model candidates. These models were tested in the industrial pilot plant during three technical campaigns. Results of the most suitable models were evaluated against to the chromatographic reference methods. Maximum relative bias of 2.88% was obtained about API target content. Absolute bias of 0.01 and 0.02% (w/w) respectively were achieved at methanol content levels of 0.10 and 0.13% (w/w). The repeatability was assessed as sufficient for the on-line monitoring of the 2 analytes. The present feasibility study confirmed the possibility to use on-line NIR spectroscopy as a PAT tool to monitor in real-time both the API and the residual methanol contents, in order to control the seeding of an API crystallization at industrial scale. Furthermore, the successful scale-up of the method proved its capability to be implemented in the manufacturing plant with the launch of the new API process. Copyright © 2013 Elsevier B.V. All rights reserved.

  10. JASPAR RESTful API: accessing JASPAR data from any programming language.

    PubMed

    Khan, Aziz; Mathelier, Anthony

    2018-05-01

    JASPAR is a widely used open-access database of curated, non-redundant transcription factor binding profiles. Currently, data from JASPAR can be retrieved as flat files or by using programming language-specific interfaces. Here, we present a programming language-independent application programming interface (API) to access JASPAR data using the Representational State Transfer (REST) architecture. The REST API enables programmatic access to JASPAR by most programming languages and returns data in eight widely used formats. Several endpoints are available to access the data and an endpoint is available to infer the TF binding profile(s) likely bound by a given DNA binding domain protein sequence. Additionally, it provides an interactive browsable interface for bioinformatics tool developers. This REST API is implemented in Python using the Django REST Framework. It is accessible at http://jaspar.genereg.net/api/ and the source code is freely available at https://bitbucket.org/CBGR/jaspar under GPL v3 license. aziz.khan@ncmm.uio.no or anthony.mathelier@ncmm.uio.no. Supplementary data are available at Bioinformatics online.

  11. PathJam: a new service for integrating biological pathway information.

    PubMed

    Glez-Peña, Daniel; Reboiro-Jato, Miguel; Domínguez, Rubén; Gómez-López, Gonzalo; Pisano, David G; Fdez-Riverola, Florentino

    2010-10-28

    Biological pathways are crucial to much of the scientific research today including the study of specific biological processes related with human diseases. PathJam is a new comprehensive and freely accessible web-server application integrating scattered human pathway annotation from several public sources. The tool has been designed for both (i) being intuitive for wet-lab users providing statistical enrichment analysis of pathway annotations and (ii) giving support to the development of new integrative pathway applications. PathJam’s unique features and advantages include interactive graphs linking pathways and genes of interest, downloadable results in fully compatible formats, GSEA compatible output files and a standardized RESTful API.

  12. Dreams deferred: Contextualizing the health and psychosocial needs of undocumented Asian and Pacific Islander young adults in Northern California.

    PubMed

    Sudhinaraset, May; Ling, Irving; To, Tu My; Melo, Jason; Quach, Thu

    2017-07-01

    There are currently 1.5 million undocumented Asians and Pacific Islanders (APIs) in the US. Undocumented API young adults, in particular, come of age in a challenging political and social climate, but little is known about their health outcomes. To our knowledge, this is the first study to assess the psychosocial needs and health status of API undocumented young adults. Guided by social capital theory, this qualitative study describes the social context of API undocumented young adults (ages 18-31), including community and government perceptions, and how social relationships influence health. This study was conducted in Northern California and included four focus group discussions (FGDs) and 24 in-depth interviews (IDIs), with 32 unique participants total. FGDs used purposeful sampling by gender (two male and two female discussions) and education status (in school and out-of-school). Findings suggest low bonding and bridging social capital. Results indicate that community distrust is high, even within the API community, due to high levels of exploitation, discrimination, and threats of deportation. Participants described how documentation status is a barrier in accessing health services, particularly mental health and sexual and reproductive health services. This study identifies trusted community groups and discusses recommendations for future research, programs, and policies. Copyright © 2017 The Authors. Published by Elsevier Ltd.. All rights reserved.

  13. Home Energy Management System - VOLTTRON Integration

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zandi, Helia

    In most Home Energy Management Systems (HEMS) available in the market, different devices running different communication protocols cannot interact with each other and exchange information. As a result of this integration, the information about different devices running different communication protocol can be accessible by other agents and devices running on VOLTTRON platform. The integration process can be used by any HEMS available in the market regardless of the programming language they use. If the existing HEMS provides an Application Programming Interface (API) based on the RESTFul architecture, that API can be used for integration. Our candidate HEMS in this projectmore » is home-assistant (Hass). An agent is implemented which can communicate with the Hass API and receives information about the devices loaded on the API. The agent publishes the information it receives on the VOLTTRON message bus so other agents can have access to this information. On the other side, for each type of devices, an agent is implemented such as Climate Agent, Lock Agent, Switch Agent, Light Agent, etc. Each of these agents is subscribed to the messages published on the message bus about their associated devices. These agents can also change the status of the devices by sending appropriate service calls to the API. Other agents and services on the platform can also access this information and coordinate their decision-making process based on this information.« less

  14. JEnsembl: a version-aware Java API to Ensembl data systems.

    PubMed

    Paterson, Trevor; Law, Andy

    2012-11-01

    The Ensembl Project provides release-specific Perl APIs for efficient high-level programmatic access to data stored in various Ensembl database schema. Although Perl scripts are perfectly suited for processing large volumes of text-based data, Perl is not ideal for developing large-scale software applications nor embedding in graphical interfaces. The provision of a novel Java API would facilitate type-safe, modular, object-orientated development of new Bioinformatics tools with which to access, analyse and visualize Ensembl data. The JEnsembl API implementation provides basic data retrieval and manipulation functionality from the Core, Compara and Variation databases for all species in Ensembl and EnsemblGenomes and is a platform for the development of a richer API to Ensembl datasources. The JEnsembl architecture uses a text-based configuration module to provide evolving, versioned mappings from database schema to code objects. A single installation of the JEnsembl API can therefore simultaneously and transparently connect to current and previous database instances (such as those in the public archive) thus facilitating better analysis repeatability and allowing 'through time' comparative analyses to be performed. Project development, released code libraries, Maven repository and documentation are hosted at SourceForge (http://jensembl.sourceforge.net).

  15. Collaborative development of predictive toxicology applications

    PubMed Central

    2010-01-01

    OpenTox provides an interoperable, standards-based Framework for the support of predictive toxicology data management, algorithms, modelling, validation and reporting. It is relevant to satisfying the chemical safety assessment requirements of the REACH legislation as it supports access to experimental data, (Quantitative) Structure-Activity Relationship models, and toxicological information through an integrating platform that adheres to regulatory requirements and OECD validation principles. Initial research defined the essential components of the Framework including the approach to data access, schema and management, use of controlled vocabularies and ontologies, architecture, web service and communications protocols, and selection and integration of algorithms for predictive modelling. OpenTox provides end-user oriented tools to non-computational specialists, risk assessors, and toxicological experts in addition to Application Programming Interfaces (APIs) for developers of new applications. OpenTox actively supports public standards for data representation, interfaces, vocabularies and ontologies, Open Source approaches to core platform components, and community-based collaboration approaches, so as to progress system interoperability goals. The OpenTox Framework includes APIs and services for compounds, datasets, features, algorithms, models, ontologies, tasks, validation, and reporting which may be combined into multiple applications satisfying a variety of different user needs. OpenTox applications are based on a set of distributed, interoperable OpenTox API-compliant REST web services. The OpenTox approach to ontology allows for efficient mapping of complementary data coming from different datasets into a unifying structure having a shared terminology and representation. Two initial OpenTox applications are presented as an illustration of the potential impact of OpenTox for high-quality and consistent structure-activity relationship modelling of REACH-relevant endpoints: ToxPredict which predicts and reports on toxicities for endpoints for an input chemical structure, and ToxCreate which builds and validates a predictive toxicity model based on an input toxicology dataset. Because of the extensible nature of the standardised Framework design, barriers of interoperability between applications and content are removed, as the user may combine data, models and validation from multiple sources in a dependable and time-effective way. PMID:20807436

  16. Collaborative development of predictive toxicology applications.

    PubMed

    Hardy, Barry; Douglas, Nicki; Helma, Christoph; Rautenberg, Micha; Jeliazkova, Nina; Jeliazkov, Vedrin; Nikolova, Ivelina; Benigni, Romualdo; Tcheremenskaia, Olga; Kramer, Stefan; Girschick, Tobias; Buchwald, Fabian; Wicker, Joerg; Karwath, Andreas; Gütlein, Martin; Maunz, Andreas; Sarimveis, Haralambos; Melagraki, Georgia; Afantitis, Antreas; Sopasakis, Pantelis; Gallagher, David; Poroikov, Vladimir; Filimonov, Dmitry; Zakharov, Alexey; Lagunin, Alexey; Gloriozova, Tatyana; Novikov, Sergey; Skvortsova, Natalia; Druzhilovsky, Dmitry; Chawla, Sunil; Ghosh, Indira; Ray, Surajit; Patel, Hitesh; Escher, Sylvia

    2010-08-31

    OpenTox provides an interoperable, standards-based Framework for the support of predictive toxicology data management, algorithms, modelling, validation and reporting. It is relevant to satisfying the chemical safety assessment requirements of the REACH legislation as it supports access to experimental data, (Quantitative) Structure-Activity Relationship models, and toxicological information through an integrating platform that adheres to regulatory requirements and OECD validation principles. Initial research defined the essential components of the Framework including the approach to data access, schema and management, use of controlled vocabularies and ontologies, architecture, web service and communications protocols, and selection and integration of algorithms for predictive modelling. OpenTox provides end-user oriented tools to non-computational specialists, risk assessors, and toxicological experts in addition to Application Programming Interfaces (APIs) for developers of new applications. OpenTox actively supports public standards for data representation, interfaces, vocabularies and ontologies, Open Source approaches to core platform components, and community-based collaboration approaches, so as to progress system interoperability goals.The OpenTox Framework includes APIs and services for compounds, datasets, features, algorithms, models, ontologies, tasks, validation, and reporting which may be combined into multiple applications satisfying a variety of different user needs. OpenTox applications are based on a set of distributed, interoperable OpenTox API-compliant REST web services. The OpenTox approach to ontology allows for efficient mapping of complementary data coming from different datasets into a unifying structure having a shared terminology and representation.Two initial OpenTox applications are presented as an illustration of the potential impact of OpenTox for high-quality and consistent structure-activity relationship modelling of REACH-relevant endpoints: ToxPredict which predicts and reports on toxicities for endpoints for an input chemical structure, and ToxCreate which builds and validates a predictive toxicity model based on an input toxicology dataset. Because of the extensible nature of the standardised Framework design, barriers of interoperability between applications and content are removed, as the user may combine data, models and validation from multiple sources in a dependable and time-effective way.

  17. HIV testing among sexually experienced Asian and Pacific Islander young women association with routine gynecologic care.

    PubMed

    Hahm, Hyeouk Chris; Song, In Han; Ozonoff, Al; Sassani, Jessica C

    2009-01-01

    To describe the proportion of HIV testing in the past 12 months among sexually experienced Asian and Pacific Islander (API) women and to investigate to what extent routine gynecologic care (RGC) increases HIV testing among API women. Data were derived from Wave III of the National Longitudinal Study of Adolescent Health (Add Health). Analyses were limited to 7,576 sexually experienced women (White, n = 4,482 [68.5%]; Black, n = 1,693 [25.6%]; Hispanic, n = 923 [13.9%]; API, n = 478 [7.2%]) aged 18-27 years. Multiple logistic regression analyses were used to estimate the association between RGC and HIV testing after controlling for predisposing, need, and enabling factors. On average, 22.8% (n = 1,504) of sexually experienced women reported HIV testing in the past year. API women had the lowest proportion of testing (17.2%), and Black women had the highest (26.2%). Overall, 60.2% of API women reported receiving RGC; however, only 15.5% of API who received RGC reported HIV testing. After controlling for covariates, significantly positive associations were found for White, Black, and Hispanic women between RGC and HIV testing; however, there was no evidence that RGC was associated with HIV testing among API women. Our data suggest that RGC does not [corrected] increase HIV testing among API women. To eliminate disparities in HIV testing service utilization among API women, appropriate efforts should be directed to better understand the barriers and facilitators of HIV testing among this population.

  18. elevatr: Access Elevation Data from Various APIs

    EPA Science Inventory

    Several web services are available that provide access to elevation data. This package provides access to several of those services and returns elevation data either as a SpatialPointsDataFrame from point elevation services or as a raster object from raster elevation services. ...

  19. The neXtProt peptide uniqueness checker: a tool for the proteomics community.

    PubMed

    Schaeffer, Mathieu; Gateau, Alain; Teixeira, Daniel; Michel, Pierre-André; Zahn-Zabal, Monique; Lane, Lydie

    2017-11-01

    The neXtProt peptide uniqueness checker allows scientists to define which peptides can be used to validate the existence of human proteins, i.e. map uniquely versus multiply to human protein sequences taking into account isobaric substitutions, alternative splicing and single amino acid variants. The pepx program is available at https://github.com/calipho-sib/pepx and can be launched from the command line or through a cgi web interface. Indexing requires a sequence file in FASTA format. The peptide uniqueness checker tool is freely available on the web at https://www.nextprot.org/tools/peptide-uniqueness-checker and from the neXtProt API at https://api.nextprot.org/. lydie.lane@sib.swiss. © The Author(s) 2017. Published by Oxford University Press.

  20. Develop 3G Application with The J2ME SATSA API

    NASA Astrophysics Data System (ADS)

    JunWu, Xu; JunLing, Liang

    This paper describes research in the use of the Security and Trust Services API for J2ME (SATSA) to develop mobile applications. for 3G networks. SATSA defines a set of APIs that allows J2ME applications to communicate with and access functionality, secure storage and cryptographic operations provided by security elements such as smart cards and Wireless Identification Modules (WIM). A Java Card application could also work as an authentication module in a J2ME-based e-bank application. The e-bank application would allow its users to access their bank accounts using their cell phones.

  1. Reactome Pengine: A web-logic API to the homo sapiens reactome.

    PubMed

    Neaves, Samuel R; Tsoka, Sophia; Millard, Louise A C

    2018-03-30

    Existing ways of accessing data from the Reactome database are limited. Either a researcher is restricted to particular queries defined by a web application programming interface (API), or they have to download the whole database. Reactome Pengine is a web service providing a logic programming based API to the human reactome. This gives researchers greater flexibility in data access than existing APIs, as users can send their own small programs (alongside queries) to Reactome Pengine. The server and an example notebook can be found at https://apps.nms.kcl.ac.uk/reactome-pengine. Source code is available at https://github.com/samwalrus/reactome-pengine and a Docker image is available at https://hub.docker.com/r/samneaves/rp4/ . samuel.neaves@kcl.ac.uk. Supplementary data are available at Bioinformatics online.

  2. How Accurately Can the Google Web Speech API Recognize and Transcribe Japanese L2 English Learners' Oral Production?

    ERIC Educational Resources Information Center

    Ashwell, Tim; Elam, Jesse R.

    2017-01-01

    The ultimate aim of our research project was to use the Google Web Speech API to automate scoring of elicited imitation (EI) tests. However, in order to achieve this goal, we had to take a number of preparatory steps. We needed to assess how accurate this speech recognition tool is in recognizing native speakers' production of the test items; we…

  3. HIV and AIDS in Suburban Asian and Pacific Islander Communities: Factors Influencing Self-Efficacy in HIV Risk Reduction

    ERIC Educational Resources Information Center

    Takahashi, Lois M.; Magalong, Michelle G.; DeBell, Paula; Fasudhani, Angela

    2006-01-01

    Though AIDS case rates among Asian Pacific Islander Americans (APIs) in the United States remain relatively low, the number has been steadily increasing. Scholars, policy makers, and service providers still know little about how confident APIs are in carrying out different HIV risk reduction strategies. This article addresses this gap by…

  4. How NASA's Atmospheric Science Data Center (ASDC) is operationally using the Esri ArcGIS Platform to improve data discoverability, accessibility and interoperability to meet the diversifying government, private, public and academic communities' driven requirements.

    NASA Astrophysics Data System (ADS)

    Tisdale, M.

    2016-12-01

    NASA's Atmospheric Science Data Center (ASDC) is operationally using the Esri ArcGIS Platform to improve data discoverability, accessibility and interoperability to meet the diversifying government, private, public and academic communities' driven requirements. The ASDC is actively working to provide their mission essential datasets as ArcGIS Image Services, Open Geospatial Consortium (OGC) Web Mapping Services (WMS), OGC Web Coverage Services (WCS) and leveraging the ArcGIS multidimensional mosaic dataset structure. Science teams and ASDC are utilizing these services, developing applications using the Web AppBuilder for ArcGIS and ArcGIS API for Javascript, and evaluating restructuring their data production and access scripts within the ArcGIS Python Toolbox framework and Geoprocessing service environment. These capabilities yield a greater usage and exposure of ASDC data holdings and provide improved geospatial analytical tools for a mission critical understanding in the areas of the earth's radiation budget, clouds, aerosols, and tropospheric chemistry.

  5. The CMS dataset bookkeeping service

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Afaq, Anzar,; /Fermilab; Dolgert, Andrew

    2007-10-01

    The CMS Dataset Bookkeeping Service (DBS) has been developed to catalog all CMS event data from Monte Carlo and Detector sources. It provides the ability to identify MC or trigger source, track data provenance, construct datasets for analysis, and discover interesting data. CMS requires processing and analysis activities at various service levels and the DBS system provides support for localized processing or private analysis, as well as global access for CMS users at large. Catalog entries can be moved among the various service levels with a simple set of migration tools, thus forming a loose federation of databases. DBS ismore » available to CMS users via a Python API, Command Line, and a Discovery web page interfaces. The system is built as a multi-tier web application with Java servlets running under Tomcat, with connections via JDBC to Oracle or MySQL database backends. Clients connect to the service through HTTP or HTTPS with authentication provided by GRID certificates and authorization through VOMS. DBS is an integral part of the overall CMS Data Management and Workflow Management systems.« less

  6. Online characterization of planetary surfaces: PlanetServer, an open-source analysis and visualization tool

    NASA Astrophysics Data System (ADS)

    Marco Figuera, R.; Pham Huu, B.; Rossi, A. P.; Minin, M.; Flahaut, J.; Halder, A.

    2018-01-01

    The lack of open-source tools for hyperspectral data visualization and analysis creates a demand for new tools. In this paper we present the new PlanetServer, a set of tools comprising a web Geographic Information System (GIS) and a recently developed Python Application Programming Interface (API) capable of visualizing and analyzing a wide variety of hyperspectral data from different planetary bodies. Current WebGIS open-source tools are evaluated in order to give an overview and contextualize how PlanetServer can help in this matters. The web client is thoroughly described as well as the datasets available in PlanetServer. Also, the Python API is described and exposed the reason of its development. Two different examples of mineral characterization of different hydrosilicates such as chlorites, prehnites and kaolinites in the Nili Fossae area on Mars are presented. As the obtained results show positive outcome in hyperspectral analysis and visualization compared to previous literature, we suggest using the PlanetServer approach for such investigations.

  7. Application Program Interface for the Orion Aerodynamics Database

    NASA Technical Reports Server (NTRS)

    Robinson, Philip E.; Thompson, James

    2013-01-01

    The Application Programming Interface (API) for the Crew Exploration Vehicle (CEV) Aerodynamic Database has been developed to provide the developers of software an easily implemented, fully self-contained method of accessing the CEV Aerodynamic Database for use in their analysis and simulation tools. The API is programmed in C and provides a series of functions to interact with the database, such as initialization, selecting various options, and calculating the aerodynamic data. No special functions (file read/write, table lookup) are required on the host system other than those included with a standard ANSI C installation. It reads one or more files of aero data tables. Previous releases of aerodynamic databases for space vehicles have only included data tables and a document of the algorithm and equations to combine them for the total aerodynamic forces and moments. This process required each software tool to have a unique implementation of the database code. Errors or omissions in the documentation, or errors in the implementation, led to a lengthy and burdensome process of having to debug each instance of the code. Additionally, input file formats differ for each space vehicle simulation tool, requiring the aero database tables to be reformatted to meet the tool s input file structure requirements. Finally, the capabilities for built-in table lookup routines vary for each simulation tool. Implementation of a new database may require an update to and verification of the table lookup routines. This may be required if the number of dimensions of a data table exceeds the capability of the simulation tools built-in lookup routines. A single software solution was created to provide an aerodynamics software model that could be integrated into other simulation and analysis tools. The highly complex Orion aerodynamics model can then be quickly included in a wide variety of tools. The API code is written in ANSI C for ease of portability to a wide variety of systems. The input data files are in standard formatted ASCII, also for improved portability. The API contains its own implementation of multidimensional table reading and lookup routines. The same aerodynamics input file can be used without modification on all implementations. The turnaround time from aerodynamics model release to a working implementation is significantly reduced

  8. Leveraging OpenStudio's Application Programming Interfaces: Preprint

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Long, N.; Ball, B.; Goldwasser, D.

    2013-11-01

    OpenStudio development efforts have been focused on providing Application Programming Interfaces (APIs) where users are able to extend OpenStudio without the need to compile the open source libraries. This paper will discuss the basic purposes and functionalities of the core libraries that have been wrapped with APIs including the Building Model, Results Processing, Advanced Analysis, UncertaintyQuantification, and Data Interoperability through Translators. Several building energy modeling applications have been produced using OpenStudio's API and Software Development Kits (SDK) including the United States Department of Energy's Asset ScoreCalculator, a mobile-based audit tool, an energy design assistance reporting protocol, and a portfolio scalemore » incentive optimization analysismethodology. Each of these software applications will be discussed briefly and will describe how the APIs were leveraged for various uses including high-level modeling, data transformations from detailed building audits, error checking/quality assurance of models, and use of high-performance computing for mass simulations.« less

  9. Visualizing the geography of genetic variants.

    PubMed

    Marcus, Joseph H; Novembre, John

    2017-02-15

    One of the key characteristics of any genetic variant is its geographic distribution. The geographic distribution can shed light on where an allele first arose, what populations it has spread to, and in turn on how migration, genetic drift, and natural selection have acted. The geographic distribution of a genetic variant can also be of great utility for medical/clinical geneticists and collectively many genetic variants can reveal population structure. Here we develop an interactive visualization tool for rapidly displaying the geographic distribution of genetic variants. Through a REST API and dynamic front-end, the Geography of Genetic Variants (GGV) browser ( http://popgen.uchicago.edu/ggv/ ) provides maps of allele frequencies in populations distributed across the globe. GGV is implemented as a website ( http://popgen.uchicago.edu/ggv/ ) which employs an API to access frequency data ( http://popgen.uchicago.edu/freq_api/ ). Python and javascript source code for the website and the API are available at: http://github.com/NovembreLab/ggv/ and http://github.com/NovembreLab/ggv-api/ . jnovembre@uchicago.edu. Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.

  10. ChEMBL web services: streamlining access to drug discovery data and utilities

    PubMed Central

    Davies, Mark; Nowotka, Michał; Papadatos, George; Dedman, Nathan; Gaulton, Anna; Atkinson, Francis; Bellis, Louisa; Overington, John P.

    2015-01-01

    ChEMBL is now a well-established resource in the fields of drug discovery and medicinal chemistry research. The ChEMBL database curates and stores standardized bioactivity, molecule, target and drug data extracted from multiple sources, including the primary medicinal chemistry literature. Programmatic access to ChEMBL data has been improved by a recent update to the ChEMBL web services (version 2.0.x, https://www.ebi.ac.uk/chembl/api/data/docs), which exposes significantly more data from the underlying database and introduces new functionality. To complement the data-focused services, a utility service (version 1.0.x, https://www.ebi.ac.uk/chembl/api/utils/docs), which provides RESTful access to commonly used cheminformatics methods, has also been concurrently developed. The ChEMBL web services can be used together or independently to build applications and data processing workflows relevant to drug discovery and chemical biology. PMID:25883136

  11. elevatr: Access Elevation Data from Various APIs | Science ...

    EPA Pesticide Factsheets

    Several web services are available that provide access to elevation data. This package provides access to several of those services and returns elevation data either as a SpatialPointsDataFrame from point elevation services or as a raster object from raster elevation services. Currently, the package supports access to the Mapzen Elevation Service, Mapzen Terrain Service, and the USGS Elevation Point Query Service. The R language for statistical computing is increasingly used for spatial data analysis . This R package, elevatr, is in response to this and provides access to elevation data from various sources directly in R. The impact of `elevatr` is that it will 1) facilitate spatial analysis in R by providing access to foundational dataset for many types of analyses (e.g. hydrology, limnology) 2) open up a new set of users and uses for APIs widely used outside of R, and 3) provide an excellent example federal open source development as promoted by the Federal Source Code Policy (https://sourcecode.cio.gov/).

  12. Quantitative analysis and comparative study of four cities green pattern in API system on the background of big data

    NASA Astrophysics Data System (ADS)

    Xin, YANG; Si-qi, WU; Qi, ZHANG

    2018-05-01

    Beijing, London, Paris, New York are typical cities in the world, so comparative study of four cities green pattern is very important to find out gap and advantage and to learn from each other. The paper will provide basis and new ideas for development of metropolises in China. On the background of big data, API (Application Programming Interface) system can provide extensive and accurate basic data to study urban green pattern in different geographical environment in domestic and foreign. On the basis of this, Average nearest neighbor tool, Kernel density tool and Standard Ellipse tool in ArcGIS platform can process and summarize data and realize quantitative analysis of green pattern. The paper summarized uniqueness of four cities green pattern and reasons of formation on basis of numerical comparison.

  13. STEREO in-situ data analysis

    NASA Astrophysics Data System (ADS)

    Schroeder, P.; Luhmann, J.; Davis, A.; Russell, C.

    STEREO s IMPACT In-situ Measurements of Particles and CME Transients investigation provides the first opportunity for long duration detailed observations of 1 AU magnetic field structures plasma and suprathermal electrons and energetic particles at points bracketing Earth s heliospheric location The PLASTIC instrument will make plasma ion composition measurements completing STEREO s comprehensive in-situ perspective Stereoscopic 3D information from the STEREO SECCHI imagers and SWAVES radio experiment will make it possible to use both multipoint and quadrature studies to connect interplanetary Coronal Mass Ejections ICME and solar wind structures to CMEs and coronal holes observed at the Sun The uniqueness of the STEREO mission requires novel data analysis tools and techniques to take advantage of the mission s full scientific potential An interactive browser with the ability to create publication-quality plots is being developed which will integrate STEREO s in-situ data with data from a variety of other missions including WIND and ACE Also an application program interface API will be provided allowing users to create custom software that ties directly into STEREO s data set The API will allow for more advanced forms of data mining than currently available through most data web services A variety of data access techniques and the development of cross-spacecraft data analysis tools will allow the larger scientific community to combine STEREO s unique in-situ data with those of other missions particularly the L1 missions and therefore to maximize

  14. ODM2 Admin Pilot Project- a Data Management Application for Observations of the Critical Zone.

    NASA Astrophysics Data System (ADS)

    Leon, M.; McDowell, W. H.; Mayorga, E.; Setiawan, L.; Hooper, R. P.

    2017-12-01

    ODM2 Admin is a tool to manage data stored in a relational database using the Observation Data Model 2 (ODM2) information model. Originally developed by the Luquillo Critical Zone Observatory (CZO) to manage a wide range of Earth observations, it has now been deployed at 6 projects: the Catalina Jemez CZO, the Dry Creek Experimental Forest, Au Sable and Manistee River sites managed by Michigan State, Tropical Response to Altered Climate Experiment (TRACE) and the Critical Zone Integrative Microbial Ecology Activity (CZIMEA) EarthCube project; most of these deployments are hosted on a Microsoft Azure cloud server managed by CUAHSI. ODM2 Admin is a web application built on the Python open-source Django framework and available for download from GitHub and DockerHub. It provides tools for data ingestion, editing, QA/QC, data visualization, browsing, mapping and documentation of equipment deployment, methods, and citations. Additional features include the ability to generate derived data values, automatically or manually create data annotations and create datasets from arbitrary groupings of results. Over 22 million time series values for more than 600 time series are being managed with ODM2 Admin across the 6 projects as well as more than 12,000 soil profiles and other measurements. ODM2 Admin links with external identifier systems through DOIs, ORCiDs and IGSNs, so cited works, details about researchers and earth sample meta-data can be accessed directly from ODM2 Admin. This application is part of a growing open source ODM2 application ecosystem under active development. ODM2 Admin can be deployed alongside other tools from the ODM2 ecosystem, including ODM2API and WOFpy, which provide access to the underlying ODM2 data through a Python API and Water One Flow web services.

  15. Development of NETCONF-Based Network Management Systems in Web Services Framework

    NASA Astrophysics Data System (ADS)

    Iijima, Tomoyuki; Kimura, Hiroyasu; Kitani, Makoto; Atarashi, Yoshifumi

    To develop a network management system (NMS) more easily, the authors developed an application programming interface (API) for configuring network devices. Because this API is used in a Java development environment, an NMS can be developed by utilizing the API and other commonly available Java libraries. It is thus possible to easily develop an NMS that is highly compatible with other IT systems. And operations that are generated from the API and that are exchanged between the NMS and network devices are based on NETCONF, which is standardized by the Internet Engineering Task Force (IETF) as a next-generation network-configuration protocol. Adopting a standardized technology ensures that the NMS developed by using the API can manage network devices provided from multi-vendors in a unified manner. Furthermore, the configuration items exchanged over NETCONF are specified in an object-oriented design. They are therefore easier to manage than such items in the Management Information Base (MIB), which is defined as data to be managed by the Simple Network Management Protocol (SNMP). We actually developed several NMSs by using the API. Evaluation of these NMSs showed that, in terms of configuration time and development time, the NMS developed by using the API performed as well as NMSs developed by using a command line interface (CLI) and SNMP. The NMS developed by using the API showed feasibility to achieve “autonomic network management” and “high interoperability with IT systems.”

  16. An adaptive software defined radio design based on a standard space telecommunication radio system API

    NASA Astrophysics Data System (ADS)

    Xiong, Wenhao; Tian, Xin; Chen, Genshe; Pham, Khanh; Blasch, Erik

    2017-05-01

    Software defined radio (SDR) has become a popular tool for the implementation and testing for communications performance. The advantage of the SDR approach includes: a re-configurable design, adaptive response to changing conditions, efficient development, and highly versatile implementation. In order to understand the benefits of SDR, the space telecommunication radio system (STRS) was proposed by NASA Glenn research center (GRC) along with the standard application program interface (API) structure. Each component of the system uses a well-defined API to communicate with other components. The benefit of standard API is to relax the platform limitation of each component for addition options. For example, the waveform generating process can support a field programmable gate array (FPGA), personal computer (PC), or an embedded system. As long as the API defines the requirements, the generated waveform selection will work with the complete system. In this paper, we demonstrate the design and development of adaptive SDR following the STRS and standard API protocol. We introduce step by step the SDR testbed system including the controlling graphic user interface (GUI), database, GNU radio hardware control, and universal software radio peripheral (USRP) tranceiving front end. In addition, a performance evaluation in shown on the effectiveness of the SDR approach for space telecommunication.

  17. JEnsembl: a version-aware Java API to Ensembl data systems

    PubMed Central

    Paterson, Trevor; Law, Andy

    2012-01-01

    Motivation: The Ensembl Project provides release-specific Perl APIs for efficient high-level programmatic access to data stored in various Ensembl database schema. Although Perl scripts are perfectly suited for processing large volumes of text-based data, Perl is not ideal for developing large-scale software applications nor embedding in graphical interfaces. The provision of a novel Java API would facilitate type-safe, modular, object-orientated development of new Bioinformatics tools with which to access, analyse and visualize Ensembl data. Results: The JEnsembl API implementation provides basic data retrieval and manipulation functionality from the Core, Compara and Variation databases for all species in Ensembl and EnsemblGenomes and is a platform for the development of a richer API to Ensembl datasources. The JEnsembl architecture uses a text-based configuration module to provide evolving, versioned mappings from database schema to code objects. A single installation of the JEnsembl API can therefore simultaneously and transparently connect to current and previous database instances (such as those in the public archive) thus facilitating better analysis repeatability and allowing ‘through time’ comparative analyses to be performed. Availability: Project development, released code libraries, Maven repository and documentation are hosted at SourceForge (http://jensembl.sourceforge.net). Contact: jensembl-develop@lists.sf.net, andy.law@roslin.ed.ac.uk, trevor.paterson@roslin.ed.ac.uk PMID:22945789

  18. Web-based network analysis and visualization using CellMaps

    PubMed Central

    Salavert, Francisco; García-Alonso, Luz; Sánchez, Rubén; Alonso, Roberto; Bleda, Marta; Medina, Ignacio; Dopazo, Joaquín

    2016-01-01

    Summary: CellMaps is an HTML5 open-source web tool that allows displaying, editing, exploring and analyzing biological networks as well as integrating metadata into them. Computations and analyses are remotely executed in high-end servers, and all the functionalities are available through RESTful web services. CellMaps can easily be integrated in any web page by using an available JavaScript API. Availability and Implementation: The application is available at: http://cellmaps.babelomics.org/ and the code can be found in: https://github.com/opencb/cell-maps. The client is implemented in JavaScript and the server in C and Java. Contact: jdopazo@cipf.es Supplementary information: Supplementary data are available at Bioinformatics online. PMID:27296979

  19. Web-based network analysis and visualization using CellMaps.

    PubMed

    Salavert, Francisco; García-Alonso, Luz; Sánchez, Rubén; Alonso, Roberto; Bleda, Marta; Medina, Ignacio; Dopazo, Joaquín

    2016-10-01

    : CellMaps is an HTML5 open-source web tool that allows displaying, editing, exploring and analyzing biological networks as well as integrating metadata into them. Computations and analyses are remotely executed in high-end servers, and all the functionalities are available through RESTful web services. CellMaps can easily be integrated in any web page by using an available JavaScript API. The application is available at: http://cellmaps.babelomics.org/ and the code can be found in: https://github.com/opencb/cell-maps The client is implemented in JavaScript and the server in C and Java. jdopazo@cipf.es Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.

  20. Subsetting and Formatting Landsat-7 LOR ETM+ and Data Products

    NASA Technical Reports Server (NTRS)

    Reid, Michael R.

    2000-01-01

    The Landsat-7 Processing System (LPS) processes Landsat-7 Enhanced Thematic Mapper (ETM+) instrument data into large, contiguous segments called "subintervals" and stores them in Level OR (LOR) data files. The LPS processed subinterval products must be subsetted and reformatted before the Level I processing systems can ingest them. The initial full subintervals produced by the LPS are stored mainly in HDF Earth Observing System (HDF-EOS) format which is an extension to the Hierarchical Data Format (HDF). The final LOR products are stored in native HDF format. Primarily the EOS Core System (ECS) and alternately the DAAC Emergency System (DES) subset the subinterval data for the operational Landsat-7 data processing systems. The HDF and HDF-EOS application programming interfaces (APIs) can be used for extensive data subsetting and data reorganization. A stand-alone subsetter tool has been developed which is based on some of the DES code. This tool makes use of the HDF and HDFEOS APIs to perform Landsat-7 LOR product subsetting and demonstrates how HDF and HDFEOS can be used for creating various configurations of full LOR products. How these APIs can be used to efficiently subset, format, and organize Landsat-7 LOR data as demonstrated by the subsetter tool and the DES is discussed.

  1. The Marine Geoscience Data System and the Global Multi-Resolution Topography Synthesis: Online Resources for Exploring Ocean Mapping Data

    NASA Astrophysics Data System (ADS)

    Ferrini, V. L.; Morton, J. J.; Carbotte, S. M.

    2016-02-01

    The Marine Geoscience Data System (MGDS: www.marine-geo.org) provides a suite of tools and services for free public access to data acquired throughout the global oceans including maps, grids, near-bottom photos, and geologic interpretations that are essential for habitat characterization and marine spatial planning. Users can explore, discover, and download data through a combination of APIs and front-end interfaces that include dynamic service-driven maps, a geospatially enabled search engine, and an easy to navigate user interface for browsing and discovering related data. MGDS offers domain-specific data curation with a team of scientists and data specialists who utilize a suite of back-end tools for introspection of data files and metadata assembly to verify data quality and ensure that data are well-documented for long-term preservation and re-use. Funded by the NSF as part of the multi-disciplinary IEDA Data Facility, MGDS also offers Data DOI registration and links between data and scientific publications. MGDS produces and curates the Global Multi-Resolution Topography Synthesis (GMRT: gmrt.marine-geo.org), a continuously updated Digital Elevation Model that seamlessly integrates multi-resolutional elevation data from a variety of sources including the GEBCO 2014 ( 1 km resolution) and International Bathymetric Chart of the Southern Ocean ( 500 m) compilations. A significant component of GMRT includes ship-based multibeam sonar data, publicly available through NOAA's National Centers for Environmental Information, that are cleaned and quality controlled by the MGDS Team and gridded at their full spatial resolution (typically 100 m resolution in the deep sea). Additional components include gridded bathymetry products contributed by individual scientists (up to meter scale resolution in places), publicly accessible regional bathymetry, and high-resolution terrestrial elevation data. New data are added to GMRT on an ongoing basis, with two scheduled releases per year. GMRT is available as both gridded data and images that can be viewed and downloaded directly through the Java application GeoMapApp (www.geomapapp.org) and the web-based GMRT MapTool. In addition, the GMRT GridServer API provides programmatic access to grids, imagery, profiles, and single point elevation values.

  2. Extending the XNAT archive tool for image and analysis management in ophthalmology research

    NASA Astrophysics Data System (ADS)

    Wahle, Andreas; Lee, Kyungmoo; Harding, Adam T.; Garvin, Mona K.; Niemeijer, Meindert; Sonka, Milan; Abràmoff, Michael D.

    2013-03-01

    In ophthalmology, various modalities and tests are utilized to obtain vital information on the eye's structure and function. For example, optical coherence tomography (OCT) is utilized to diagnose, screen, and aid treatment of eye diseases like macular degeneration or glaucoma. Such data are complemented by photographic retinal fundus images and functional tests on the visual field. DICOM isn't widely used yet, though, and frequently images are encoded in proprietary formats. The eXtensible Neuroimaging Archive Tool (XNAT) is an open-source NIH-funded framework for research PACS and is in use at the University of Iowa for neurological research applications. Its use for ophthalmology was hence desirable but posed new challenges due to data types thus far not considered and the lack of standardized formats. We developed custom tools for data types not natively recognized by XNAT itself using XNAT's low-level REST API. Vendor-provided tools can be included as necessary to convert proprietary data sets into valid DICOM. Clients can access the data in a standardized format while still retaining the original format if needed by specific analysis tools. With respective project-specific permissions, results like segmentations or quantitative evaluations can be stored as additional resources to previously uploaded datasets. Applications can use our abstract-level Python or C/C++ API to communicate with the XNAT instance. This paper describes concepts and details of the designed upload script templates, which can be customized to the needs of specific projects, and the novel client-side communication API which allows integration into new or existing research applications.

  3. Armenian Virtual Observatory: Services and Data Sharing

    NASA Astrophysics Data System (ADS)

    Knyazyan, A. V.; Astsatryan, H. V.; Mickaelian, A. M.

    2016-06-01

    The main aim of this article is to introduce the data management and services of the Armenian Virtual Observatory (ArVO), which consists of user friendly data management mechanisms, a new and productive cross-correlation service, and data sharing API based on international standards and protocols.

  4. Targeted Expansion Project for Outreach and Treatment for Substance Abuse and HIV Risk Behaviors in Asian and Pacific Islander Communities

    ERIC Educational Resources Information Center

    Nemoto, Tooru; Iwamoto, Mariko; Kamitani, Emiko; Morris, Anne; Sakata, Maria

    2011-01-01

    Access to culturally competent HIV/AIDS and substance abuse treatment and prevention services is limited for Asian and Pacific Islanders (APIs). Based on the intake data for a community outreach project in the San Francisco Bay Area (N = 1,349), HIV risk behaviors were described among the targeted API risk groups. The self-reported HIV prevalence…

  5. The jmzQuantML programming interface and validator for the mzQuantML data standard.

    PubMed

    Qi, Da; Krishna, Ritesh; Jones, Andrew R

    2014-03-01

    The mzQuantML standard from the HUPO Proteomics Standards Initiative has recently been released, capturing quantitative data about peptides and proteins, following analysis of MS data. We present a Java application programming interface (API) for mzQuantML called jmzQuantML. The API provides robust bridges between Java classes and elements in mzQuantML files and allows random access to any part of the file. The API provides read and write capabilities, and is designed to be embedded in other software packages, enabling mzQuantML support to be added to proteomics software tools (http://code.google.com/p/jmzquantml/). The mzQuantML standard is designed around a multilevel validation system to ensure that files are structurally and semantically correct for different proteomics quantitative techniques. In this article, we also describe a Java software tool (http://code.google.com/p/mzquantml-validator/) for validating mzQuantML files, which is a formal part of the data standard. © 2014 The Authors. Proteomics published by Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  6. Towards Monitoring-as-a-service for Scientific Computing Cloud applications using the ElasticSearch ecosystem

    NASA Astrophysics Data System (ADS)

    Bagnasco, S.; Berzano, D.; Guarise, A.; Lusso, S.; Masera, M.; Vallero, S.

    2015-12-01

    The INFN computing centre in Torino hosts a private Cloud, which is managed with the OpenNebula cloud controller. The infrastructure offers Infrastructure-as-a-Service (IaaS) and Platform-as-a-Service (PaaS) services to different scientific computing applications. The main stakeholders of the facility are a grid Tier-2 site for the ALICE collaboration at LHC, an interactive analysis facility for the same experiment and a grid Tier-2 site for the BESIII collaboration, plus an increasing number of other small tenants. The dynamic allocation of resources to tenants is partially automated. This feature requires detailed monitoring and accounting of the resource usage. We set up a monitoring framework to inspect the site activities both in terms of IaaS and applications running on the hosted virtual instances. For this purpose we used the ElasticSearch, Logstash and Kibana (ELK) stack. The infrastructure relies on a MySQL database back-end for data preservation and to ensure flexibility to choose a different monitoring solution if needed. The heterogeneous accounting information is transferred from the database to the ElasticSearch engine via a custom Logstash plugin. Each use-case is indexed separately in ElasticSearch and we setup a set of Kibana dashboards with pre-defined queries in order to monitor the relevant information in each case. For the IaaS metering, we developed sensors for the OpenNebula API. The IaaS level information gathered through the API is sent to the MySQL database through an ad-hoc developed RESTful web service. Moreover, we have developed a billing system for our private Cloud, which relies on the RabbitMQ message queue for asynchronous communication to the database and on the ELK stack for its graphical interface. The Italian Grid accounting framework is also migrating to a similar set-up. Concerning the application level, we used the Root plugin TProofMonSenderSQL to collect accounting data from the interactive analysis facility. The BESIII virtual instances used to be monitored with Zabbix, as a proof of concept we also retrieve the information contained in the Zabbix database. In this way we have achieved a uniform monitoring interface for both the IaaS and the scientific applications, mostly leveraging off-the-shelf tools. At present, we are working to define a model for monitoring-as-a-service, based on the tools described above, which the Cloud tenants can easily configure to suit their specific needs.

  7. BioAssay Research Database (BARD): chemical biology and probe-development enabled by structured metadata and result types

    PubMed Central

    Howe, E.A.; de Souza, A.; Lahr, D.L.; Chatwin, S.; Montgomery, P.; Alexander, B.R.; Nguyen, D.-T.; Cruz, Y.; Stonich, D.A.; Walzer, G.; Rose, J.T.; Picard, S.C.; Liu, Z.; Rose, J.N.; Xiang, X.; Asiedu, J.; Durkin, D.; Levine, J.; Yang, J.J.; Schürer, S.C.; Braisted, J.C.; Southall, N.; Southern, M.R.; Chung, T.D.Y.; Brudz, S.; Tanega, C.; Schreiber, S.L.; Bittker, J.A.; Guha, R.; Clemons, P.A.

    2015-01-01

    BARD, the BioAssay Research Database (https://bard.nih.gov/) is a public database and suite of tools developed to provide access to bioassay data produced by the NIH Molecular Libraries Program (MLP). Data from 631 MLP projects were migrated to a new structured vocabulary designed to capture bioassay data in a formalized manner, with particular emphasis placed on the description of assay protocols. New data can be submitted to BARD with a user-friendly set of tools that assist in the creation of appropriately formatted datasets and assay definitions. Data published through the BARD application program interface (API) can be accessed by researchers using web-based query tools or a desktop client. Third-party developers wishing to create new tools can use the API to produce stand-alone tools or new plug-ins that can be integrated into BARD. The entire BARD suite of tools therefore supports three classes of researcher: those who wish to publish data, those who wish to mine data for testable hypotheses, and those in the developer community who wish to build tools that leverage this carefully curated chemical biology resource. PMID:25477388

  8. Characterization of solid polymer dispersions of active pharmaceutical ingredients by 19F MAS NMR and factor analysis

    NASA Astrophysics Data System (ADS)

    Urbanova, Martina; Brus, Jiri; Sedenkova, Ivana; Policianova, Olivia; Kobera, Libor

    In this contribution the ability of 19F MAS NMR spectroscopy to probe structural variability of poorly water-soluble drugs formulated as solid dispersions in polymer matrices is discussed. The application potentiality of the proposed approach is demonstrated on a moderately sized active pharmaceutical ingredient (API, Atorvastatin) exhibiting extensive polymorphism. In this respect, a range of model systems with the API incorporated in the matrix of polvinylpyrrolidone (PVP) was prepared. The extent of mixing of both components was determined by T1(1H) and T1ρ(1H) relaxation experiments, and it was found that the API forms nanosized domains. Subsequently it was found out that the polymer matrix induces two kinds of changes in 19F MAS NMR spectra. At first, this is a high-frequency shift reaching 2-3 ppm which is independent on molecular structure of the API and which results from the long-range polarization of the electron cloud around 19F nucleus induced by electrostatic fields of the polymer matrix. At second, this is broadening of the signals and formation of shoulders reflecting changes in molecular arrangement of the API. To avoid misleading in the interpretation of the recorded 19F MAS NMR spectra, because both the contributions act simultaneously, we applied chemometric approach based on multivariate analysis. It is demonstrated that factor analysis of the recorded spectra can separate both these spectral contributions, and the subtle structural differences in the molecular arrangement of the API in the nanosized domains can be traced. In this way 19F MAS NMR spectra of both pure APIs and APIs in solid dispersions can be directly compared. The proposed strategy thus provides a powerful tool for the analysis of new formulations of fluorinated pharmaceutical substances in polymer matrices.

  9. Manufacturing Amorphous Solid Dispersions with a Tailored Amount of Crystallized API for Biopharmaceutical Testing.

    PubMed

    Theil, Frank; Milsmann, Johanna; Anantharaman, Sankaran; van Lishaut, Holger

    2018-05-07

    The preparation of an amorphous solid dispersion (ASD) by dissolving a poorly water-soluble active pharmaceutical ingredient (API) in a polymer matrix can improve the bioavailability by orders of magnitude. Crystallization of the API in the ASD, though, is an inherent threat for bioavailability. Commonly, the impact of crystalline API on the drug release of the dosage form is studied with samples containing spiked crystallinity. These spiked samples possess implicit differences compared to native crystalline samples, regarding size and spatial distribution of the crystals as well as their molecular environment. In this study, we demonstrate that it is possible to grow defined amounts of crystalline API in solid dosage forms, which enables us to study the biopharmaceutical impact of actual crystallization. For this purpose, we studied the crystal growth in fenofibrate tablets over time under an elevated moisture using transmission Raman spectroscopy (TRS). As a nondestructive method to assess API crystallinity in ASD formulations, TRS enables the monitoring of crystal growth in individual dosage forms. Once the kinetic trace of the crystal growth for a certain environmental condition is determined, this method can be used to produce samples with defined amounts of crystallized API. To investigate the biopharmaceutical impact of crystallized API, non-QC dissolution methods were used, designed to identify differences between the various amounts of crystalline materials present. The drug release in the samples manufactured in this fashion was compared to that of samples with spiked crystallinity. In this study, we present for the first time a method for targeted crystallization of amorphous tablets to simulate crystallized ASDs. This methodology is a valuable tool to generate model systems for biopharmaceutical studies on the impact of crystallinity on the bioavailability.

  10. DOE Office of Scientific and Technical Information (OSTI.GOV)

    The open source Project Haystack initiative defines meta data and communication standards related to data from buildings and intelligent devices. The Project Haystack REST API defines standard formats and operations for exchanging Haystack tagged data over HTTP. The HaystackRuby gem wraps calls to this REST API to enable Ruby application to easily integrate data hosted on a Project Haystack compliant server. The HaystackRuby gem was developed at the National Renewable Energy Lab to support applications related to campus energy. We hope that this tool may be useful to others.

  11. Terahertz Tools Advance Imaging for Security, Industry

    NASA Technical Reports Server (NTRS)

    2010-01-01

    Picometrix, a wholly owned subsidiary of Advanced Photonix Inc. (API), of Ann Arbor, Michigan, invented the world s first commercial terahertz system. The company improved the portability and capabilities of their systems through Small Business Innovation Research (SBIR) agreements with Langley Research Center to provide terahertz imaging capabilities for inspecting the space shuttle external tanks and orbiters. Now API s systems make use of the unique imaging capacity of terahertz radiation on manufacturing floors, for thickness measurements of coatings, pharmaceutical tablet production, and even art conservation.

  12. Apis - a Digital Inventory of Archaeological Heritage Based on Remote Sensing Data

    NASA Astrophysics Data System (ADS)

    Doneus, M.; Forwagner, U.; Liem, J.; Sevara, C.

    2017-08-01

    Heritage managers are in need of dynamic spatial inventories of archaeological and cultural heritage that provide them with multipurpose tools to interactively understand information about archaeological heritage within its landscape context. Specifically, linking site information with the respective non-invasive prospection data is of increasing importance as it allows for the assessment of inherent uncertainties related to the use and interpretation of remote sensing data by the educated and knowledgeable heritage manager. APIS, the archaeological prospection information system of the Aerial Archive of the University of Vienna, is specifically designed to meet these needs. It provides storage and easy access to all data concerning aerial photographs and archaeological sites through a single GIS-based application. Furthermore, APIS has been developed in an open source environment, which allows it to be freely distributed and modified. This combination in one single open source system facilitates an easy workflow for data management, interpretation, storage, and retrieval. APIS and a sample dataset will be released free of charge under creative commons license in near future.

  13. Evolution of a mini-scale biphasic dissolution model: Impact of model parameters on partitioning of dissolved API and modelling of in vivo-relevant kinetics.

    PubMed

    Locher, Kathrin; Borghardt, Jens M; Frank, Kerstin J; Kloft, Charlotte; Wagner, Karl G

    2016-08-01

    Biphasic dissolution models are proposed to have good predictive power for the in vivo absorption. The aim of this study was to improve our previously introduced mini-scale dissolution model to mimic in vivo situations more realistically and to increase the robustness of the experimental model. Six dissolved APIs (BCS II) were tested applying the improved mini-scale biphasic dissolution model (miBIdi-pH-II). The influence of experimental model parameters including various excipients, API concentrations, dual paddle and its rotation speed was investigated. The kinetics in the biphasic model was described applying a one- and four-compartment pharmacokinetic (PK) model. The improved biphasic dissolution model was robust related to differing APIs and excipient concentrations. The dual paddle guaranteed homogenous mixing in both phases; the optimal rotation speed was 25 and 75rpm for the aqueous and the octanol phase, respectively. A one-compartment PK model adequately characterised the data of fully dissolved APIs. A four-compartment PK model best quantified dissolution, precipitation, and partitioning also of undissolved amounts due to realistic pH profiles. The improved dissolution model is a powerful tool for investigating the interplay between dissolution, precipitation and partitioning of various poorly soluble APIs (BCS II). In vivo-relevant PK parameters could be estimated applying respective PK models. Copyright © 2016 Elsevier B.V. All rights reserved.

  14. Using the Generic Mapping Tools From Within the MATLAB, Octave and Julia Computing Environments

    NASA Astrophysics Data System (ADS)

    Luis, J. M. F.; Wessel, P.

    2016-12-01

    The Generic Mapping Tools (GMT) is a widely used software infrastructure tool set for analyzing and displaying geoscience data. Its power to analyze and process data and produce publication-quality graphics has made it one of several standard processing toolsets used by a large segment of the Earth and Ocean Sciences. GMT's strengths lie in superior publication-quality vector graphics, geodetic-quality map projections, robust data processing algorithms scalable to enormous data sets, and ability to run under all common operating systems. The GMT tool chest offers over 120 modules sharing a common set of command options, file structures, and documentation. GMT modules are command line tools that accept input and write output, and this design allows users to write scripts in which one module's output becomes another module's input, creating highly customized GMT workflows. With the release of GMT 5, these modules are high-level functions with a C API, potentially allowing users access to high-level GMT capabilities from any programmable environment. Many scientists who use GMT also use other computational tools, such as MATLAB® and its clone Octave. We have built a MATLAB/Octave interface on top of the GMT 5 C API. Thus, MATLAB or Octave now has full access to all GMT modules as well as fundamental input/output of GMT data objects via a MEX function. Internally, the GMT/MATLAB C API defines six high-level composite data objects that handle input and output of data via individual GMT modules. These are data tables, grids, text tables (text/data mixed records), color palette tables, raster images (1-4 color bands), and PostScript. The API is responsible for translating between the six GMT objects and the corresponding native MATLAB objects. References to data arrays are passed if transposing of matrices is not required. The GMT and MATLAB/Octave combination is extremely flexible, letting the user harvest the general numerical and graphical capabilities of both systems, and represents a giant step forward in interoperability between GMT and other software package. We will present examples of the symbiotic benefits of combining these platforms. Two other extensions are also in the works: a nearly finished Julia wrapper and an embryonic Python module. Publication supported by FCT- project UID/GEO/50019/2013 - Instituto D. Luiz

  15. ChEMBL web services: streamlining access to drug discovery data and utilities.

    PubMed

    Davies, Mark; Nowotka, Michał; Papadatos, George; Dedman, Nathan; Gaulton, Anna; Atkinson, Francis; Bellis, Louisa; Overington, John P

    2015-07-01

    ChEMBL is now a well-established resource in the fields of drug discovery and medicinal chemistry research. The ChEMBL database curates and stores standardized bioactivity, molecule, target and drug data extracted from multiple sources, including the primary medicinal chemistry literature. Programmatic access to ChEMBL data has been improved by a recent update to the ChEMBL web services (version 2.0.x, https://www.ebi.ac.uk/chembl/api/data/docs), which exposes significantly more data from the underlying database and introduces new functionality. To complement the data-focused services, a utility service (version 1.0.x, https://www.ebi.ac.uk/chembl/api/utils/docs), which provides RESTful access to commonly used cheminformatics methods, has also been concurrently developed. The ChEMBL web services can be used together or independently to build applications and data processing workflows relevant to drug discovery and chemical biology. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.

  16. A Data Services Upgrade for Advanced Composition Explorer (ACE) Data

    NASA Astrophysics Data System (ADS)

    Davis, A. J.; Hamell, G.

    2008-12-01

    Since early in 1998, NASA's Advanced Composition Explorer (ACE) spacecraft has provided continuous measurements of solar wind, interplanetary magnetic field, and energetic particle activity from L1, located approximately 0.01 AU sunward of Earth. The spacecraft has enough fuel to stay in orbit about L1 until ~2024. The ACE Science Center (ASC) provides access to ACE data, and performs level 1 and browse data processing for the science instruments. Thanks to a NASA Data Services Upgrade grant, we have recently retooled our legacy web interface to ACE data, enhancing data subsetting capabilities and improving online plotting options. We have also integrated a new application programming interface (API) and we are working to ensure that it will be compatible with emerging Virtual Observatory (VO) data services standards. The new API makes extensive use of metadata created using the Space Physics Archive Search and Extract (SPASE) data model. We describe these recent improvements to the ACE Science Center data services, and our plans for integrating these services into the VO system.

  17. Activity-Centric Approach to Distributed Programming

    NASA Technical Reports Server (NTRS)

    Levy, Renato; Satapathy, Goutam; Lang, Jun

    2004-01-01

    The first phase of an effort to develop a NASA version of the Cybele software system has been completed. To give meaning to even a highly abbreviated summary of the modifications to be embodied in the NASA version, it is necessary to present the following background information on Cybele: Cybele is a proprietary software infrastructure for use by programmers in developing agent-based application programs [complex application programs that contain autonomous, interacting components (agents)]. Cybele provides support for event handling from multiple sources, multithreading, concurrency control, migration, and load balancing. A Cybele agent follows a programming paradigm, called activity-centric programming, that enables an abstraction over system-level thread mechanisms. Activity centric programming relieves application programmers of the complex tasks of thread management, concurrency control, and event management. In order to provide such functionality, activity-centric programming demands support of other layers of software. This concludes the background information. In the first phase of the present development, a new architecture for Cybele was defined. In this architecture, Cybele follows a modular service-based approach to coupling of the programming and service layers of software architecture. In a service-based approach, the functionalities supported by activity-centric programming are apportioned, according to their characteristics, among several groups called services. A well-defined interface among all such services serves as a path that facilitates the maintenance and enhancement of such services without adverse effect on the whole software framework. The activity-centric application-program interface (API) is part of a kernel. The kernel API calls the services by use of their published interface. This approach makes it possible for any application code written exclusively under the API to be portable for any configuration of Cybele.

  18. A General Provincial Situation Visualization System Based on iPhone Operating System of Shandong Province

    NASA Astrophysics Data System (ADS)

    Ye, Z.; Xiang, H.

    2014-04-01

    The paper discusses the basic principles and the problem solutions during the design and implementation of the mobile GIS system, and base on the research result, we developed the General Provincial Situation Visualization System Based on iOS of Shandong Province. The system is developed in the Objective-C programming language, and use the ArcGIS Runtime SDK for IOS as the development tool to call the "World-map Shandong" services to implement the development of the General Provincial Situation Visualization System Based on iOS devices. The system is currently available for download in the Appstore and is chosen as the typical application case of ESRI China ArcGIS API for iOS.

  19. Environmental Models as a Service: Enabling Interoperability through RESTful Endpoints and API Documentation

    EPA Science Inventory

    Achieving interoperability in environmental modeling has evolved as software technology has progressed. The recent rise of cloud computing and proliferation of web services initiated a new stage for creating interoperable systems. Scientific programmers increasingly take advantag...

  20. Environmental Models as a Service: Enabling Interoperability through RESTful Endpoints and API Documentation.

    EPA Science Inventory

    Achieving interoperability in environmental modeling has evolved as software technology has progressed. The recent rise of cloud computing and proliferation of web services initiated a new stage for creating interoperable systems. Scientific programmers increasingly take advantag...

  1. STEREO In-situ Data Analysis

    NASA Astrophysics Data System (ADS)

    Schroeder, P. C.; Luhmann, J. G.; Davis, A. J.; Russell, C. T.

    2007-05-01

    STEREO's IMPACT (In-situ Measurements of Particles and CME Transients) investigation provides the first opportunity for long duration, detailed observations of 1 AU magnetic field structures, plasma and suprathermal electrons, and energetic particles at points bracketing Earth's heliospheric location. The PLASTIC instrument takes plasma ion composition measurements completing STEREO's comprehensive in-situ perspective. Stereoscopic/3D information from the STEREO SECCHI imagers and SWAVES radio experiment make it possible to use both multipoint and quadrature studies to connect interplanetary Coronal Mass Ejections (ICME) and solar wind structures to CMEs and coronal holes observed at the Sun. The uniqueness of the STEREO mission requires novel data analysis tools and techniques to take advantage of the mission's full scientific potential. An interactive browser with the ability to create publication-quality plots has been developed which integrates STEREO's in-situ data with data from a variety of other missions including WIND and ACE. Static summary plots and a key-parameter type data set with a related online browser provide alternative data access. Finally, an application program interface (API) is provided allowing users to create custom software that ties directly into STEREO's data set. The API allows for more advanced forms of data mining than currently available through most web-based data services. A variety of data access techniques and the development of cross- spacecraft data analysis tools allow the larger scientific community to combine STEREO's unique in-situ data with those of other missions, particularly the L1 missions, and, therefore, to maximize STEREO's scientific potential in gaining a greater understanding of the heliosphere.

  2. Alternative Fuels Data Center: Case Studies

    Science.gov Websites

    goal of saving money, reducing its environmental impact, and increasing employees' satisfaction. Learn Legislation Data & Tools Widgets Data Downloads APIs About Project Assistance News & Features Spanish

  3. Secure and Resilient Cloud Computing for the Department of Defense

    DTIC Science & Technology

    2015-11-16

    platform as a service (PaaS), and software as a service ( SaaS )—that target system administrators, developers, and end-users respectively (see Table 2...interfaces (API) and services Medium Amazon Elastic MapReduce, MathWorks Cloud, Red Hat OpenShift SaaS Full-fledged applications Low Google gMail

  4. GOTCHA

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Poliakoff, David; Legendre, Matt

    2017-03-29

    GOTCHA is a runtime API intercepting function calls between shared libraries. It is intended to be used by HPC Tools (i.e., performance analysis tools like Open/SpeedShop, HPCToolkit, TAU, etc.). 2:18 PMThese other tools can use Gotch to intercept interesting functions, such as MPI functions, and collect performance metrics about those functions. We intend for this to be open-source software that gets adopted by other open-s0urse tools that are used at LLNL.

  5. SeqLib: a C ++ API for rapid BAM manipulation, sequence alignment and sequence assembly

    PubMed Central

    Wala, Jeremiah; Beroukhim, Rameen

    2017-01-01

    Abstract We present SeqLib, a C ++ API and command line tool that provides a rapid and user-friendly interface to BAM/SAM/CRAM files, global sequence alignment operations and sequence assembly. Four C libraries perform core operations in SeqLib: HTSlib for BAM access, BWA-MEM and BLAT for sequence alignment and Fermi for error correction and sequence assembly. Benchmarking indicates that SeqLib has lower CPU and memory requirements than leading C ++ sequence analysis APIs. We demonstrate an example of how minimal SeqLib code can extract, error-correct and assemble reads from a CRAM file and then align with BWA-MEM. SeqLib also provides additional capabilities, including chromosome-aware interval queries and read plotting. Command line tools are available for performing integrated error correction, micro-assemblies and alignment. Availability and Implementation: SeqLib is available on Linux and OSX for the C ++98 standard and later at github.com/walaj/SeqLib. SeqLib is released under the Apache2 license. Additional capabilities for BLAT alignment are available under the BLAT license. Contact: jwala@broadinstitue.org; rameen@broadinstitute.org PMID:28011768

  6. SeqLib: a C ++ API for rapid BAM manipulation, sequence alignment and sequence assembly.

    PubMed

    Wala, Jeremiah; Beroukhim, Rameen

    2017-03-01

    We present SeqLib, a C ++ API and command line tool that provides a rapid and user-friendly interface to BAM/SAM/CRAM files, global sequence alignment operations and sequence assembly. Four C libraries perform core operations in SeqLib: HTSlib for BAM access, BWA-MEM and BLAT for sequence alignment and Fermi for error correction and sequence assembly. Benchmarking indicates that SeqLib has lower CPU and memory requirements than leading C ++ sequence analysis APIs. We demonstrate an example of how minimal SeqLib code can extract, error-correct and assemble reads from a CRAM file and then align with BWA-MEM. SeqLib also provides additional capabilities, including chromosome-aware interval queries and read plotting. Command line tools are available for performing integrated error correction, micro-assemblies and alignment. SeqLib is available on Linux and OSX for the C ++98 standard and later at github.com/walaj/SeqLib. SeqLib is released under the Apache2 license. Additional capabilities for BLAT alignment are available under the BLAT license. jwala@broadinstitue.org ; rameen@broadinstitute.org. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com

  7. Environmental Models as a Service: Enabling Interoperability through RESTful Endpoints and API Documentation (presentation)

    EPA Science Inventory

    Achieving interoperability in environmental modeling has evolved as software technology has progressed. The recent rise of cloud computing and proliferation of web services initiated a new stage for creating interoperable systems. Scientific programmers increasingly take advantag...

  8. Gee Fu: a sequence version and web-services database tool for genomic assembly, genome feature and NGS data.

    PubMed

    Ramirez-Gonzalez, Ricardo; Caccamo, Mario; MacLean, Daniel

    2011-10-01

    Scientists now use high-throughput sequencing technologies and short-read assembly methods to create draft genome assemblies in just days. Tools and pipelines like the assembler, and the workflow management environments make it easy for a non-specialist to implement complicated pipelines to produce genome assemblies and annotations very quickly. Such accessibility results in a proliferation of assemblies and associated files, often for many organisms. These assemblies get used as a working reference by lots of different workers, from a bioinformatician doing gene prediction or a bench scientist designing primers for PCR. Here we describe Gee Fu, a database tool for genomic assembly and feature data, including next-generation sequence alignments. Gee Fu is an instance of a Ruby-On-Rails web application on a feature database that provides web and console interfaces for input, visualization of feature data via AnnoJ, access to data through a web-service interface, an API for direct data access by Ruby scripts and access to feature data stored in BAM files. Gee Fu provides a platform for storing and sharing different versions of an assembly and associated features that can be accessed and updated by bench biologists and bioinformaticians in ways that are easy and useful for each. http://tinyurl.com/geefu dan.maclean@tsl.ac.uk.

  9. SeqHound: biological sequence and structure database as a platform for bioinformatics research

    PubMed Central

    2002-01-01

    Background SeqHound has been developed as an integrated biological sequence, taxonomy, annotation and 3-D structure database system. It provides a high-performance server platform for bioinformatics research in a locally-hosted environment. Results SeqHound is based on the National Center for Biotechnology Information data model and programming tools. It offers daily updated contents of all Entrez sequence databases in addition to 3-D structural data and information about sequence redundancies, sequence neighbours, taxonomy, complete genomes, functional annotation including Gene Ontology terms and literature links to PubMed. SeqHound is accessible via a web server through a Perl, C or C++ remote API or an optimized local API. It provides functionality necessary to retrieve specialized subsets of sequences, structures and structural domains. Sequences may be retrieved in FASTA, GenBank, ASN.1 and XML formats. Structures are available in ASN.1, XML and PDB formats. Emphasis has been placed on complete genomes, taxonomy, domain and functional annotation as well as 3-D structural functionality in the API, while fielded text indexing functionality remains under development. SeqHound also offers a streamlined WWW interface for simple web-user queries. Conclusions The system has proven useful in several published bioinformatics projects such as the BIND database and offers a cost-effective infrastructure for research. SeqHound will continue to develop and be provided as a service of the Blueprint Initiative at the Samuel Lunenfeld Research Institute. The source code and examples are available under the terms of the GNU public license at the Sourceforge site http://sourceforge.net/projects/slritools/ in the SLRI Toolkit. PMID:12401134

  10. Ground System Architectures Workshop GMSEC SERVICES SUITE (GSS): an Agile Development Story

    NASA Technical Reports Server (NTRS)

    Ly, Vuong

    2017-01-01

    The GMSEC (Goddard Mission Services Evolution Center) Services Suite (GSS) is a collection of tools and software services along with a robust customizable web-based portal that enables the user to capture, monitor, report, and analyze system-wide GMSEC data. Given our plug-and-play architecture and the needs for rapid system development, we opted to follow the Scrum Agile Methodology for software development. Being one of the first few projects to implement the Agile methodology at NASA GSFC, in this presentation we will present our approaches, tools, successes, and challenges in implementing this methodology. The GMSEC architecture provides a scalable, extensible ground and flight system for existing and future missions. GMSEC comes with a robust Application Programming Interface (GMSEC API) and a core set of Java-based GMSEC components that facilitate the development of a GMSEC-based ground system. Over the past few years, we have seen an upbeat in the number of customers who are moving from a native desktop application environment to a web based environment particularly for data monitoring and analysis. We also see a need to provide separation of the business logic from the GUI display for our Java-based components and also to consolidate all the GUI displays into one interface. This combination of separation and consolidation brings immediate value to a GMSEC-based ground system through increased ease of data access via a uniform interface, built-in security measures, centralized configuration management, and ease of feature extensibility.

  11. Targeted expansion project for outreach and treatment for substance abuse and HIV risk behaviors in Asian and Pacific Islander communities.

    PubMed

    Nemoto, Tooru; Iwamoto, Mariko; Kamitani, Emiko; Morris, Anne; Sakata, Maria

    2011-04-01

    Access to culturally competent HIV/AIDS and substance abuse treatment and prevention services is limited for Asian and Pacific Islanders (APIs). Based on the intake data for a community outreach project in the San Francisco Bay Area (N = 1,349), HIV risk behaviors were described among the targeted API risk groups. The self-reported HIV prevalence was 6% among MSM. Inconsistent condom use for vaginal sex with casual partners in the past 6 months was reported among substance users (43%) and incarcerated participants (60%), whereas 26% of men who have sex with men reported inconsistent condom use for anal sex with casual partners. Overall, 56% and 29% had engaged in sex with casual partners under the influence of alcohol and drugs in the past 6 months, respectively. Although API organizations in the Bay Area have spearheaded HIV prevention, future programs must address substance use issues in relation to sexual risk behaviors, specific to API risk groups.

  12. The Joy of Playing with Oceanographic Data

    NASA Astrophysics Data System (ADS)

    Smith, A. T.; Xing, Z.; Armstrong, E. M.; Thompson, C. K.; Huang, T.

    2013-12-01

    The web is no longer just an after thought. It is no longer just a presentation layer filled with HTML, CSS, JavaScript, Frameworks, 3D, and more. It has become the medium of our communication. It is the database of all databases. It is the computing platform of all platforms. It has transformed the way we do science. Web service is the de facto method for communication between machines over the web. Representational State Transfer (REST) has standardized the way we architect services and their interfaces. In the Earth Science domain, we are familiar with tools and services such as Open-Source Project for Network Data Access Protocol (OPeNDAP), Thematic Realtime Environmental Distributed Data Services (THREDDS), and Live Access Server (LAS). We are also familiar with various data formats such as NetCDF3/4, HDF4/5, GRIB, TIFF, etc. One of the challenges for the Earth Science community is accessing information within these data. There are community-accepted readers that our users can download and install. However, the Application Programming Interface (API) between these readers is not standardized, which leads to non-portable applications. Webification (w10n) is an emerging technology, developed at the Jet Propulsion Laboratory, which exploits the hierarchical nature of a science data artifact to assign a URL to each element within the artifact. (e.g. a granule file). By embracing standards such as JSON, XML, and HTML5 and predictable URL, w10n provides a simple interface that enables tool-builders and researchers to develop portable tools/applications to interact with artifacts of various formats. The NASA Physical Oceanographic Distributed Active Archive Center (PO.DAAC) is the designated data center for observational products relevant to the physical state of the ocean. Over the past year PO.DAAC has been evaluating w10n technology by webifying its archive holdings to provide simplified access to oceanographic science artifacts and as a service to enable future tools and services development. In this talk, we will focus on a w10n-based system called Distributed Oceanographic Webification Service (DOWS) being developed at PO.DAAC to provide a newer and simpler method for working with observational data artifacts. As a continued effort at PO.DAAC to provide better tools and services to visualize our data, the talk will discuss the latest in web-based data visualization tools/frameworks (such as d3.js, Three.js, Leaflet.js, and more) and techniques for working with webified oceanographic science data in both a 2D and 3D web approach.

  13. RELEASE NOTES FOR MODELS-3 VERSION 4.1 PATCH: SMOKE TOOL AND FILE CONVERTER

    EPA Science Inventory

    This software patch to the Models-3 system corrects minor errors in the Models-3 framework, provides substantial improvements in the ASCII to I/O API format conversion of the File Converter utility, and new functionalities for the SMOKE Tool. Version 4.1 of the Models-3 system...

  14. Design and Implementation of Surrounding Transaction Plotting and Management System Based on Google Map API

    NASA Astrophysics Data System (ADS)

    Cao, Y. B.; Hua, Y. X.; Zhao, J. X.; Guo, S. M.

    2013-11-01

    With China's rapid economic development and comprehensive national strength growing, Border work has become a long-term and important task in China's diplomatic work. How to implement rapid plotting, real-time sharing and mapping surrounding affairs has taken great significance for government policy makers and diplomatic staff. However, at present the already exists Boundary information system are mainly have problems of Geospatial data update is heavily workload, plotting tools are in a state of serious lack of, Geographic events are difficult to share, this phenomenon has seriously hampered the smooth development of the border task. The development and progress of Geographic information system technology especially the development of Web GIS offers the possibility to solve the above problems, this paper adopts four layers of B/S architecture, with the support of Google maps service, uses the free API which is offered by Google maps and its features of openness, ease of use, sharing characteristics, highresolution images to design and implement the surrounding transaction plotting and management system based on the web development technology of ASP.NET, C#, Ajax. The system can provide decision support for government policy makers as well as diplomatic staff's real-time plotting and sharing of surrounding information. The practice has proved that the system has good usability and strong real-time.

  15. The development, validity, and reliability of the Addiction Profile Index (API).

    PubMed

    Ögel, Kültegin; Evren, Cüneyt; Karadağ, Figen; Gürol, Defne Tamar

    2012-01-01

    The objective of this study was to develop a practical questionnaire for multidimensional assessment of problems associated with alcohol and substance abuse that would also be useful for treatment planning. The Addiction Profile Index (API) is a self-report questionnaire consisting of 37 items and the following 5 subscales: characteristics of substance use; dependency diagnosis; the effects of subsance use on the user; craving; motivation to quit using substances. The study included 345 alcohol and/or substance abusers from 2 addiction treatment clinics and a prison addiction service. The validity of the questionnaire was assessed using the Michigan Alcoholism Screening Test (MAST), Readiness to Change Questionnaire (SOCRATES), Penn Alcohol Craving Scale (PACS), Drug Craving Scale (DCS), Structured Clinical Interview for DSM-IV Axis I Disorders (SCID-I), and Addiction Severity Index (ASI). The Cronbach's alpha coefficient for the total API was 0.89 and for the subscales it ranged from 0.63 to 0.86. Item-total correlation coefficients ranged from 0.42 to 0.89. The Spearman Brown split-half method coefficient for the total API was 0.83. In all, 4 factors were obtained using explanatory factor analysis that represented 52.3% of the total variance. The API craving subscale was observed to be consistent with PACS and the API motivation subscale was consistent with SOCRATES. The API total score was strongly correlated with the mean MAST score, and the composite ASI medical status, substance use, legal status, and family social relations subscale scores. Based on ROC analyses, the area under curve was 0.90. With a total API cut-off score of 4, the scale's sensitivity and specificity 0.85 was 0.78, respectively. The findings show that the API is a valid and reliable questionnaire that can be used to measure the severity of different dimensions of substance dependency.

  16. Chronic toxicity of an environmentally relevant mixture of pharmaceuticals to three aquatic organisms (alga, daphnid, and fish).

    PubMed

    Watanabe, Haruna; Tamura, Ikumi; Abe, Ryoko; Takanobu, Hitomi; Nakamura, Ataru; Suzuki, Toshinari; Hirose, Akihiko; Nishimura, Tetsuji; Tatarazako, Norihisa

    2016-04-01

    Principles of concentration addition and independent action have been used as effective tools to predict mixture toxicity based on individual component toxicity. The authors investigated the toxicity of a pharmaceutical mixture composed of the top 10 detected active pharmaceutical ingredients (APIs) in the Tama River (Tokyo, Japan) in a relevant concentration ratio. Both individual and mixture toxicities of the 10 APIs were evaluated by 3 short-term chronic toxicity tests using the alga Pseudokirchneriella subcapitata, the daphnid Ceriodaphnia dubia, and the zebrafish Danio rerio. With the exception of clarithromycin toxicity to alga, the no-observed-effect concentration of individual APIs for each test species was dramatically higher than the highest concentration of APIs found in the environment. The mixture of 10 APIs resulted in toxicity to alga, daphnid, and fish at 6.25 times, 100 times, and 15,000 times higher concentrations, respectively, than the environmental concentrations of individual APIs. Predictions by concentration addition and independent action were nearly identical for alga, as clarithromycin was the predominant toxicant in the mixture. Both predictions described the observed mixture toxicity to alga fairly well, whereas they slightly underestimated the observed mixture toxicity in the daphnid test. In the fish embryo test, the observed toxicity fell between the predicted toxicity by concentration addition and independent action. These results suggested that the toxicity of environmentally relevant pharmaceutical mixtures could be predicted by individual toxicity using either concentration addition or independent action. © 2015 SETAC.

  17. NASA GSFC Space Weather Center - Innovative Space Weather Dissemination: Web-Interfaces, Mobile Applications, and More

    NASA Technical Reports Server (NTRS)

    Maddox, Marlo; Zheng, Yihua; Rastaetter, Lutz; Taktakishvili, A.; Mays, M. L.; Kuznetsova, M.; Lee, Hyesook; Chulaki, Anna; Hesse, Michael; Mullinix, Richard; hide

    2012-01-01

    The NASA GSFC Space Weather Center (http://swc.gsfc.nasa.gov) is committed to providing forecasts, alerts, research, and educational support to address NASA's space weather needs - in addition to the needs of the general space weather community. We provide a host of services including spacecraft anomaly resolution, historical impact analysis, real-time monitoring and forecasting, custom space weather alerts and products, weekly summaries and reports, and most recently - video casts. There are many challenges in providing accurate descriptions of past, present, and expected space weather events - and the Space Weather Center at NASA GSFC employs several innovative solutions to provide access to a comprehensive collection of both observational data, as well as space weather model/simulation data. We'll describe the challenges we've faced with managing hundreds of data streams, running models in real-time, data storage, and data dissemination. We'll also highlight several systems and tools that are utilized by the Space Weather Center in our daily operations, all of which are available to the general community as well. These systems and services include a web-based application called the Integrated Space Weather Analysis System (iSWA http://iswa.gsfc.nasa.gov), two mobile space weather applications for both IOS and Android devices, an external API for web-service style access to data, google earth compatible data products, and a downloadable client-based visualization tool.

  18. Alternative Fuels Data Center: News and Features

    Science.gov Websites

    ; Features Spanish Resources Contacts News and Features The Alternative Fuels Data Center (AFDC) helps Legislation Data & Tools Widgets Data Downloads APIs About Project Assistance News & Features Spanish

  19. Alternative Fuels Data Center: Disclaimer

    Science.gov Websites

    fueling infrastructure data to create a comprehensive tool for consumers, fleets, and DOE stakeholders APIs About Project Assistance News & Features Spanish Resources Contacts The AFDC is a resource of

  20. The Semantic Automated Discovery and Integration (SADI) Web service Design-Pattern, API and Reference Implementation

    PubMed Central

    2011-01-01

    Background The complexity and inter-related nature of biological data poses a difficult challenge for data and tool integration. There has been a proliferation of interoperability standards and projects over the past decade, none of which has been widely adopted by the bioinformatics community. Recent attempts have focused on the use of semantics to assist integration, and Semantic Web technologies are being welcomed by this community. Description SADI - Semantic Automated Discovery and Integration - is a lightweight set of fully standards-compliant Semantic Web service design patterns that simplify the publication of services of the type commonly found in bioinformatics and other scientific domains. Using Semantic Web technologies at every level of the Web services "stack", SADI services consume and produce instances of OWL Classes following a small number of very straightforward best-practices. In addition, we provide codebases that support these best-practices, and plug-in tools to popular developer and client software that dramatically simplify deployment of services by providers, and the discovery and utilization of those services by their consumers. Conclusions SADI Services are fully compliant with, and utilize only foundational Web standards; are simple to create and maintain for service providers; and can be discovered and utilized in a very intuitive way by biologist end-users. In addition, the SADI design patterns significantly improve the ability of software to automatically discover appropriate services based on user-needs, and automatically chain these into complex analytical workflows. We show that, when resources are exposed through SADI, data compliant with a given ontological model can be automatically gathered, or generated, from these distributed, non-coordinating resources - a behaviour we have not observed in any other Semantic system. Finally, we show that, using SADI, data dynamically generated from Web services can be explored in a manner very similar to data housed in static triple-stores, thus facilitating the intersection of Web services and Semantic Web technologies. PMID:22024447

  1. The Semantic Automated Discovery and Integration (SADI) Web service Design-Pattern, API and Reference Implementation.

    PubMed

    Wilkinson, Mark D; Vandervalk, Benjamin; McCarthy, Luke

    2011-10-24

    The complexity and inter-related nature of biological data poses a difficult challenge for data and tool integration. There has been a proliferation of interoperability standards and projects over the past decade, none of which has been widely adopted by the bioinformatics community. Recent attempts have focused on the use of semantics to assist integration, and Semantic Web technologies are being welcomed by this community. SADI - Semantic Automated Discovery and Integration - is a lightweight set of fully standards-compliant Semantic Web service design patterns that simplify the publication of services of the type commonly found in bioinformatics and other scientific domains. Using Semantic Web technologies at every level of the Web services "stack", SADI services consume and produce instances of OWL Classes following a small number of very straightforward best-practices. In addition, we provide codebases that support these best-practices, and plug-in tools to popular developer and client software that dramatically simplify deployment of services by providers, and the discovery and utilization of those services by their consumers. SADI Services are fully compliant with, and utilize only foundational Web standards; are simple to create and maintain for service providers; and can be discovered and utilized in a very intuitive way by biologist end-users. In addition, the SADI design patterns significantly improve the ability of software to automatically discover appropriate services based on user-needs, and automatically chain these into complex analytical workflows. We show that, when resources are exposed through SADI, data compliant with a given ontological model can be automatically gathered, or generated, from these distributed, non-coordinating resources - a behaviour we have not observed in any other Semantic system. Finally, we show that, using SADI, data dynamically generated from Web services can be explored in a manner very similar to data housed in static triple-stores, thus facilitating the intersection of Web services and Semantic Web technologies.

  2. Eldercare Locator

    MedlinePlus

    ... About Resources Welcome to the Eldercare Locator, a public service of the U.S. Administration on Aging connecting you to services for older ... Modified: 11/16/2016 1:05:04 PM Administration on Aging Feedback | Admin Portal | API ... FOIA | Plain Writing | No Fear Act | ACL.gov | USA.gov | HHS.gov

  3. The Snow Data System at NASA JPL

    NASA Astrophysics Data System (ADS)

    Laidlaw, R.; Painter, T. H.; Mattmann, C. A.; Ramirez, P.; Bormann, K.; Brodzik, M. J.; Burgess, A. B.; Rittger, K.; Goodale, C. E.; Joyce, M.; McGibbney, L. J.; Zimdars, P.

    2014-12-01

    NASA JPL's Snow Data System has a data-processing pipeline powered by Apache OODT, an open source software tool. The pipeline has been running for several years and has successfully generated a significant amount of cryosphere data, including MODIS-based products such as MODSCAG, MODDRFS and MODICE, with historical and near-real time windows and covering regions such as the Artic, Western US, Alaska, Central Europe, Asia, South America, Australia and New Zealand. The team continues to improve the pipeline, using monitoring tools such as Ganglia to give an overview of operations, and improving fault-tolerance with automated recovery scripts. Several alternative adaptations of the Snow Covered Area and Grain size (SCAG) algorithm are being investigated. These include using VIIRS and Landsat TM/ETM+ satellite data as inputs. Parallel computing techniques are being considered for core SCAG processing, such as using the PyCUDA Python API to utilize multi-core GPU architectures. An experimental version of MODSCAG is also being developed for the Google Earth Engine platform, a cloud-based service.

  4. Integration of active pharmaceutical ingredient solid form selection and particle engineering into drug product design.

    PubMed

    Ticehurst, Martyn David; Marziano, Ivan

    2015-06-01

    This review seeks to offer a broad perspective that encompasses an understanding of the drug product attributes affected by active pharmaceutical ingredient (API) physical properties, their link to solid form selection and the role of particle engineering. While the crucial role of active pharmaceutical ingredient (API) solid form selection is universally acknowledged in the pharmaceutical industry, the value of increasing effort to understanding the link between solid form, API physical properties and drug product formulation and manufacture is now also being recognised. A truly holistic strategy for drug product development should focus on connecting solid form selection, particle engineering and formulation design to both exploit opportunities to access simpler manufacturing operations and prevent failures. Modelling and predictive tools that assist in establishing these links early in product development are discussed. In addition, the potential for differences between the ingoing API physical properties and those in the final product caused by drug product processing is considered. The focus of this review is on oral solid dosage forms and dry powder inhaler products for lung delivery. © 2015 Royal Pharmaceutical Society.

  5. The prevalence of the honeybee brood pathogens Ascosphaera apis, Paenibacillus larvae and Melissococcus plutonius in Spanish apiaries determined with a new multiplex PCR assay

    PubMed Central

    Garrido-Bailón, Encarna; Higes, Mariano; Martínez-Salvador, Amparo; Antúnez, Karina; Botías, Cristina; Meana, Aránzazu; Prieto, Lourdes; Martín-Hernández, Raquel

    2013-01-01

    The microorganisms Ascosphaera apis, Paenibacillus larvae and Melissococcus plutonius are the three most important pathogens that affect honeybee brood. The aim of the present study was to evaluate the prevalence of these pathogens in honeybee colonies and to elucidate their role in the honeybee colony losses in Spain. In order to get it, a multiplex polymerase chain reaction (PCR) assay was developed to simultaneously amplify the16S ribosomal ribonucleic acid (rRNA) gene of P. larvae and M. plutonius, and the 5.8S rRNA gene of A. apis. The multiplex PCR assay provides a quick and specific tool that successfully detected the three infectious pathogens (P. larvae, M. plutonius and A. apis) in brood and adult honeybee samples without the need for microbiological culture. This technique was then used to evaluate the prevalence of these pathogens in Spanish honeybee colonies in 2006 and 2007, revealing our results a low prevalence of these pathogens in most of the geographic areas studied. PMID:23919248

  6. Genetics Home Reference: prolidase deficiency

    MedlinePlus

    ... mutations as a tool to investigate structure-function relationship. J Hum Genet. 2004;49(9):500-6. ... for Links Data Files & API Site Map Subscribe Customer Support USA.gov Copyright Privacy Accessibility FOIA Viewers & ...

  7. When Will It Be... USNO Seasons and Apsides Calculator

    NASA Astrophysics Data System (ADS)

    Chizek Frouard, Malynda; Bartlett, Jennifer Lynn

    2018-01-01

    The turning of the Earth’s seasons (solstices and equinoxes) and apsides (perihelions and aphelions) are times often used in observational astronomy and also of interest to the public. To avoid tedious calculations, the U.S. Naval Observatory (USNO) has developed an on-line interactive calculator, Earth’s Seasons and Apsides to provide information about events between 1600 and 2200. The new data service uses an Application Programming Interface (API), which returns values in JavaScript Object Notation (JSON) that can be incorporated into third-party websites or applications. For a requested year, the Earth’s Seasons and Apsides API provides the Gregorian calendar date and time of the Vernal Equinox, Summer Solstice, Autumnal Equinox, Winter Solstice, Aphelion, and Perihelion. The user may specify the time zone for their results, including the optional addition of U.S. daylight saving time for years after 1966.On-line documentation for using the API-enabled Earth’s Seasons and Apsides is available, including sample calls (http://aa.usno.navy.mil/data/docs/api.php). A traditional forms-based interface is available as well (http://aa.usno.navy.mil/data/docs/EarthSeasons.php). This data service replaces the popular Earth's Seasons: Equinoxes, Solstices, Perihelion, and Aphelion page that provided a static list of events for 2000–2025. The USNO also provides API-enabled data services for Complete Sun and Moon Data for One Day (http://aa.usno.navy.mil/data/docs/RS_OneDay.php), Dates of the Primary Phases of the Moon (http://aa.usno.navy.mil/data/docs/MoonPhase.php), Selected Christian Observances (http://aa.usno.navy.mil/data/docs/easter.php), Selected Islamic Observances (http://aa.usno.navy.mil/data/docs/islamic.php), Selected Jewish Observances (http://aa.usno.navy.mil/data/docs/passover.php), Julian Date Conversion (http://aa.usno.navy.mil/data/docs/JulianDate.php), and Sidereal Time (http://aa.usno.navy.mil/data/docs/siderealtime.php) as well as its Solar Eclipse Computer (http://aa.usno.navy.mil/data/docs/SolarEclipses.php).

  8. DAVID-WS: a stateful web service to facilitate gene/protein list analysis

    PubMed Central

    Jiao, Xiaoli; Sherman, Brad T.; Huang, Da Wei; Stephens, Robert; Baseler, Michael W.; Lane, H. Clifford; Lempicki, Richard A.

    2012-01-01

    Summary: The database for annotation, visualization and integrated discovery (DAVID), which can be freely accessed at http://david.abcc.ncifcrf.gov/, is a web-based online bioinformatics resource that aims to provide tools for the functional interpretation of large lists of genes/proteins. It has been used by researchers from more than 5000 institutes worldwide, with a daily submission rate of ∼1200 gene lists from ∼400 unique researchers, and has been cited by more than 6000 scientific publications. However, the current web interface does not support programmatic access to DAVID, and the uniform resource locator (URL)-based application programming interface (API) has a limit on URL size and is stateless in nature as it uses URL request and response messages to communicate with the server, without keeping any state-related details. DAVID-WS (web service) has been developed to automate user tasks by providing stateful web services to access DAVID programmatically without the need for human interactions. Availability: The web service and sample clients (written in Java, Perl, Python and Matlab) are made freely available under the DAVID License at http://david.abcc.ncifcrf.gov/content.jsp?file=WS.html. Contact: xiaoli.jiao@nih.gov; rlempicki@nih.gov PMID:22543366

  9. DAVID-WS: a stateful web service to facilitate gene/protein list analysis.

    PubMed

    Jiao, Xiaoli; Sherman, Brad T; Huang, Da Wei; Stephens, Robert; Baseler, Michael W; Lane, H Clifford; Lempicki, Richard A

    2012-07-01

    The database for annotation, visualization and integrated discovery (DAVID), which can be freely accessed at http://david.abcc.ncifcrf.gov/, is a web-based online bioinformatics resource that aims to provide tools for the functional interpretation of large lists of genes/proteins. It has been used by researchers from more than 5000 institutes worldwide, with a daily submission rate of ∼1200 gene lists from ∼400 unique researchers, and has been cited by more than 6000 scientific publications. However, the current web interface does not support programmatic access to DAVID, and the uniform resource locator (URL)-based application programming interface (API) has a limit on URL size and is stateless in nature as it uses URL request and response messages to communicate with the server, without keeping any state-related details. DAVID-WS (web service) has been developed to automate user tasks by providing stateful web services to access DAVID programmatically without the need for human interactions. The web service and sample clients (written in Java, Perl, Python and Matlab) are made freely available under the DAVID License at http://david.abcc.ncifcrf.gov/content.jsp?file=WS.html.

  10. Hadoop-BAM: directly manipulating next generation sequencing data in the cloud.

    PubMed

    Niemenmaa, Matti; Kallio, Aleksi; Schumacher, André; Klemelä, Petri; Korpelainen, Eija; Heljanko, Keijo

    2012-03-15

    Hadoop-BAM is a novel library for the scalable manipulation of aligned next-generation sequencing data in the Hadoop distributed computing framework. It acts as an integration layer between analysis applications and BAM files that are processed using Hadoop. Hadoop-BAM solves the issues related to BAM data access by presenting a convenient API for implementing map and reduce functions that can directly operate on BAM records. It builds on top of the Picard SAM JDK, so tools that rely on the Picard API are expected to be easily convertible to support large-scale distributed processing. In this article we demonstrate the use of Hadoop-BAM by building a coverage summarizing tool for the Chipster genome browser. Our results show that Hadoop offers good scalability, and one should avoid moving data in and out of Hadoop between analysis steps.

  11. Cloud-Based Speech Technology for Assistive Technology Applications (CloudCAST).

    PubMed

    Cunningham, Stuart; Green, Phil; Christensen, Heidi; Atria, José Joaquín; Coy, André; Malavasi, Massimiliano; Desideri, Lorenzo; Rudzicz, Frank

    2017-01-01

    The CloudCAST platform provides a series of speech recognition services that can be integrated into assistive technology applications. The platform and the services provided by the public API are described. Several exemplar applications have been developed to demonstrate the platform to potential developers and users.

  12. HIV and AIDS in suburban Asian and Pacific Islander communities: factors influencing self-efficacy in HIV risk reduction.

    PubMed

    Takahashi, Lois M; Magalong, Michelle G; Debell, Paula; Fasudhani, Angela

    2006-12-01

    Though AIDS case rates among Asian Pacific Islander Americans (APIs) in the United States remain relatively low, the number has been steadily increasing. Scholars, policy makers, and service providers still know little about how confident APIs are in carrying out different HIV risk reduction strategies. This article addresses this gap by presenting an analysis of a survey of API women and youth in Orange County, California (N = 313), a suburban county in southern California with large concentrations of Asian residents. Multivariate logistic regression models using subsamples of API women and API youth respondents were used. Variations in reported self-efficacy for female respondents were explained by acculturation, comfort in asking medical practitioners about HIV/AIDS, and to a lesser degree, education, household size, whether respondents were currently dating, HIV knowledge, and whether respondents believed that HIV could be identified by physical appearance. For respondents younger than 25 years, variations in self-efficacy were related to gender, age, acculturation, HIV knowledge, taking-over-the-counter medicines for illness, whether respondents were dating, and to a lesser degree, employment, recent serious illness, whether they believe that one could identify HIV by how one looks, and believing that illness was caused by germs. Implications for HIV prevention programs and future research are provided.

  13. Genome sequencing and comparative genomics of honey bee microsporidia, Nosema apis reveal novel insights into host-parasite interactions.

    PubMed

    Chen, Yan ping; Pettis, Jeffery S; Zhao, Yan; Liu, Xinyue; Tallon, Luke J; Sadzewicz, Lisa D; Li, Renhua; Zheng, Huoqing; Huang, Shaokang; Zhang, Xuan; Hamilton, Michele C; Pernal, Stephen F; Melathopoulos, Andony P; Yan, Xianghe; Evans, Jay D

    2013-07-05

    The microsporidia parasite Nosema contributes to the steep global decline of honey bees that are critical pollinators of food crops. There are two species of Nosema that have been found to infect honey bees, Nosema apis and N. ceranae. Genome sequencing of N. apis and comparative genome analysis with N. ceranae, a fully sequenced microsporidia species, reveal novel insights into host-parasite interactions underlying the parasite infections. We applied the whole-genome shotgun sequencing approach to sequence and assemble the genome of N. apis which has an estimated size of 8.5 Mbp. We predicted 2,771 protein- coding genes and predicted the function of each putative protein using the Gene Ontology. The comparative genomic analysis led to identification of 1,356 orthologs that are conserved between the two Nosema species and genes that are unique characteristics of the individual species, thereby providing a list of virulence factors and new genetic tools for studying host-parasite interactions. We also identified a highly abundant motif in the upstream promoter regions of N. apis genes. This motif is also conserved in N. ceranae and other microsporidia species and likely plays a role in gene regulation across the microsporidia. The availability of the N. apis genome sequence is a significant addition to the rapidly expanding body of microsprodian genomic data which has been improving our understanding of eukaryotic genome diversity and evolution in a broad sense. The predicted virulent genes and transcriptional regulatory elements are potential targets for innovative therapeutics to break down the life cycle of the parasite.

  14. The Myths of Chinese Images Revisited: Persuasive Strategies in Hong Kong Anti-Drug Public Service Announcements.

    ERIC Educational Resources Information Center

    Wong, Wendy Siuyi; Cuklanz, Lisa M.

    Hong Kong's Department of Information Services has been producing and systematically airing public service announcements (known as announcements in the public interest, or APIs) on a variety of selected themes such as cleaning up Hong Kong, road safety, crime, drugs, and health issues for over 20 years. The television announcements are shown every…

  15. High-throughput molecular identification of Staphylococcus spp. isolated from a clean room facility in an environmental monitoring program

    PubMed Central

    2010-01-01

    Background The staphylococci are one of the most common environmental isolates found in clean room facility. Consequently, isolation followed by comprehensive and accurate identification is an essential step in any environmental monitoring program. Findings We have used the API Staph identification kit (bioMérieux, France) which depends on the expression of metabolic activities and or morphological features to identify the Staphylococcus isolates. The API staphylococci showed low sensitivity in the identification of some species, so we performed molecular methods based on PCR based fingerprinting of glyceraldehyde-3-phosphate dehydrogenase encoding gene as useful taxonomic tool for examining Staphylococcus isolates. Conclusions Our results showed that PCR protocol used in this study which depends on genotypic features was relatively accurate, rapid, sensitive and superior in the identification of at least 7 species of Staphylococcus than API Staph which depends on phenotypic features. PMID:21047438

  16. High-throughput molecular identification of Staphylococcus spp. isolated from a clean room facility in an environmental monitoring program.

    PubMed

    Sheraba, Norhan S; Yassin, Aymen S; Amin, Magdy A

    2010-11-04

    The staphylococci are one of the most common environmental isolates found in clean room facility. Consequently, isolation followed by comprehensive and accurate identification is an essential step in any environmental monitoring program. We have used the API Staph identification kit (bioMérieux, France) which depends on the expression of metabolic activities and or morphological features to identify the Staphylococcus isolates. The API staphylococci showed low sensitivity in the identification of some species, so we performed molecular methods based on PCR based fingerprinting of glyceraldehyde-3-phosphate dehydrogenase encoding gene as useful taxonomic tool for examining Staphylococcus isolates. Our results showed that PCR protocol used in this study which depends on genotypic features was relatively accurate, rapid, sensitive and superior in the identification of at least 7 species of Staphylococcus than API Staph which depends on phenotypic features.

  17. Design of a Model Execution Framework: Repetitive Object-Oriented Simulation Environment (ROSE)

    NASA Technical Reports Server (NTRS)

    Gray, Justin S.; Briggs, Jeffery L.

    2008-01-01

    The ROSE framework was designed to facilitate complex system analyses. It completely divorces the model execution process from the model itself. By doing so ROSE frees the modeler to develop a library of standard modeling processes such as Design of Experiments, optimizers, parameter studies, and sensitivity studies which can then be applied to any of their available models. The ROSE framework accomplishes this by means of a well defined API and object structure. Both the API and object structure are presented here with enough detail to implement ROSE in any object-oriented language or modeling tool.

  18. SiteDB: Marshalling people and resources available to CMS

    NASA Astrophysics Data System (ADS)

    Metson, S.; Bonacorsi, D.; Dias Ferreira, M.; Egeland, R.

    2010-04-01

    In a collaboration the size of CMS (approx. 3000 users, and almost 100 computing centres of varying size) communication and accurate information about the sites it has access to is vital in co-ordinating the multitude of computing tasks required for smooth running. SiteDB is a tool developed by CMS to track sites available to the collaboration, the allocation to CMS of resources available at those sites and the associations between CMS members and the sites (as either a manager/operator of the site or a member of a group associated to the site). It is used to track the roles a person has for an associated site or group. SiteDB eases the coordination load for the operations teams by providing a consistent interface to manage communication with the people working at a site, by identifying who is responsible for a given task or service at a site and by offering a uniform interface to information on CMS contacts and sites. SiteDB provides api's and reports for other CMS tools to use to access the information it contains, for instance enabling CRAB to use "user friendly" names when black/white listing CE's, providing role based authentication and authorisation for other web based services and populating various troubleshooting squads in external ticketing systems in use daily by CMS Computing operations.

  19. TH-D-BRB-04: Pinnacle Scripting: Improving Efficiency While Maintaining Safety

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Moore, J.

    2016-06-15

    Scripting capabilities and application programming interfaces (APIs) are becoming commonly available in modern treatment planning systems. These links to the treatment planning system (TPS) allow users to read data from the TPS, and in some cases use TPS functionality and write data back to the TPS. Such tools are powerful extensions, allowing automation of routine clinical tasks and supporting research, particularly research involving repetitive tasks on large patient populations. The data and functionality exposed by scripting/API capabilities is vendor dependent, as are the languages used by script/API engines, such as the Microsoft .NET framework or Python. Scripts deployed in amore » clinical environment must be commissioned and validated like any other software tool. This session will provide an overview of scripting applications and a discussion of best practices, followed by a practical introduction to the scripting capabilities of three commercial treatment planning systems. Learning Objectives: Understand the scripting capabilities available in several treatment planning systems Learn how to get started using scripting capabilities Understand the best practices for safe script deployment in a clinical environment R. Popple, Varian Medical Systems has provided research support unrelated to the topic of this session.R. Cardan, Varian Medical Systems for grant research, product evaluation, and teaching honorarium.« less

  20. Engineering and algorithm design for an image processing Api: a technical report on ITK--the Insight Toolkit.

    PubMed

    Yoo, Terry S; Ackerman, Michael J; Lorensen, William E; Schroeder, Will; Chalana, Vikram; Aylward, Stephen; Metaxas, Dimitris; Whitaker, Ross

    2002-01-01

    We present the detailed planning and execution of the Insight Toolkit (ITK), an application programmers interface (API) for the segmentation and registration of medical image data. This public resource has been developed through the NLM Visible Human Project, and is in beta test as an open-source software offering under cost-free licensing. The toolkit concentrates on 3D medical data segmentation and registration algorithms, multimodal and multiresolution capabilities, and portable platform independent support for Windows, Linux/Unix systems. This toolkit was built using current practices in software engineering. Specifically, we embraced the concept of generic programming during the development of these tools, working extensively with C++ templates and the freedom and flexibility they allow. Software development tools for distributed consortium-based code development have been created and are also publicly available. We discuss our assumptions, design decisions, and some lessons learned.

  1. GLobal Integrated Design Environment (GLIDE): A Concurrent Engineering Application

    NASA Technical Reports Server (NTRS)

    McGuire, Melissa L.; Kunkel, Matthew R.; Smith, David A.

    2010-01-01

    The GLobal Integrated Design Environment (GLIDE) is a client-server software application purpose-built to mitigate issues associated with real time data sharing in concurrent engineering environments and to facilitate discipline-to-discipline interaction between multiple engineers and researchers. GLIDE is implemented in multiple programming languages utilizing standardized web protocols to enable secure parameter data sharing between engineers and researchers across the Internet in closed and/or widely distributed working environments. A well defined, HyperText Transfer Protocol (HTTP) based Application Programming Interface (API) to the GLIDE client/server environment enables users to interact with GLIDE, and each other, within common and familiar tools. One such common tool, Microsoft Excel (Microsoft Corporation), paired with its add-in API for GLIDE, is discussed in this paper. The top-level examples given demonstrate how this interface improves the efficiency of the design process of a concurrent engineering study while reducing potential errors associated with manually sharing information between study participants.

  2. HammerCloud: A Stress Testing System for Distributed Analysis

    NASA Astrophysics Data System (ADS)

    van der Ster, Daniel C.; Elmsheuser, Johannes; Úbeda García, Mario; Paladin, Massimo

    2011-12-01

    Distributed analysis of LHC data is an I/O-intensive activity which places large demands on the internal network, storage, and local disks at remote computing facilities. Commissioning and maintaining a site to provide an efficient distributed analysis service is therefore a challenge which can be aided by tools to help evaluate a variety of infrastructure designs and configurations. HammerCloud is one such tool; it is a stress testing service which is used by central operations teams, regional coordinators, and local site admins to (a) submit arbitrary number of analysis jobs to a number of sites, (b) maintain at a steady-state a predefined number of jobs running at the sites under test, (c) produce web-based reports summarizing the efficiency and performance of the sites under test, and (d) present a web-interface for historical test results to both evaluate progress and compare sites. HammerCloud was built around the distributed analysis framework Ganga, exploiting its API for grid job management. HammerCloud has been employed by the ATLAS experiment for continuous testing of many sites worldwide, and also during large scale computing challenges such as STEP'09 and UAT'09, where the scale of the tests exceeded 10,000 concurrently running and 1,000,000 total jobs over multi-day periods. In addition, HammerCloud is being adopted by the CMS experiment; the plugin structure of HammerCloud allows the execution of CMS jobs using their official tool (CRAB).

  3. Open Core Data: Connecting scientific drilling data to scientists and community data resources

    NASA Astrophysics Data System (ADS)

    Fils, D.; Noren, A. J.; Lehnert, K.; Diver, P.

    2016-12-01

    Open Core Data (OCD) is an innovative, efficient, and scalable infrastructure for data generated by scientific drilling and coring to improve discoverability, accessibility, citability, and preservation of data from the oceans and continents. OCD is building on existing community data resources that manage, store, publish, and preserve scientific drilling data, filling a critical void that currently prevents linkages between these and other data systems and tools to realize the full potential of data generated through drilling and coring. We are developing this functionality through Linked Open Data (LOD) and semantic patterns that enable data access through the use of community ontologies such as GeoLink (geolink.org, an EarthCube Building Block), a collection of protocols, formats and vocabularies from a set of participating geoscience repositories. Common shared concepts of classes such as cruise, dataset, person and others allow easier resolution of common references through shared resource IDs. These graphs are then made available via SPARQL as well as incorporated into web pages following schema.org approaches. Additionally the W3C PROV vocabulary is under evaluation for use for documentation of provenance. Further, the application of persistent identifiers for samples (IGSNs); datasets, expeditions, and projects (DOIs); and people (ORCIDs), combined with LOD approaches, provides methods to resolve and incorporate metadata and datasets. Application Program Interfaces (APIs) complement these semantic approaches to the OCD data holdings. APIs are exposed following the Swagger guidelines (swagger.io) and will be evolved into the OpenAPI (openapis.org) approach. Currently APIs are in development for the NSF funded Flyover Country mobile geoscience app (fc.umn.edu), the Neotoma Paleoecology Database (neotomadb.org), Magnetics Information Consortium (MagIC; earthref.org/MagIC), and other community tools and data systems, as well as for internal OCD use.

  4. Mission Services Evolution Center Message Bus

    NASA Technical Reports Server (NTRS)

    Mayorga, Arturo; Bristow, John O.; Butschky, Mike

    2011-01-01

    The Goddard Mission Services Evolution Center (GMSEC) Message Bus is a robust, lightweight, fault-tolerant middleware implementation that supports all messaging capabilities of the GMSEC API. This architecture is a distributed software system that routes messages based on message subject names and knowledge of the locations in the network of the interested software components.

  5. GIS tool to locate major Sikh temples in USA

    NASA Astrophysics Data System (ADS)

    Sharma, Saumya

    This tool is a GIS based interactive and graphical user interface tool, which locates the major Sikh temples of USA on a map. This tool is using Java programming language along with MOJO (Map Object Java Object) provided by ESRI that is the organization that provides the GIS software. It also includes some of the integration with Google's API's like Google Translator API. This application will tell users about the origin of Sikhism in India and USA, the major Sikh temples in each state of USA, location, name and detail information through their website. The primary purpose of this application is to make people aware about this religion and culture. This tool will also measure the distance between two temple points in a map and display the result in miles and kilometers. Also, there is an added support to convert each temple's website language from English to Punjabi or any other language using a language convertor tool so that people from different nationalities can understand their culture. By clicking on each point on a map, a new window will pop up showing the picture of the temple and a hyperlink that will redirect to the website of that particular temple .It will also contain links to their dance, music, history, and also a help menu to guide the users to use the software efficiently.

  6. The Materials Data Facility: Data Services to Advance Materials Science Research

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Blaiszik, B.; Chard, K.; Pruyne, J.

    2016-07-06

    With increasingly strict data management requirements from funding agencies and institutions, expanding focus on the challenges of research replicability, and growing data sizes and heterogeneity, new data needs are emerging in the materials community. The materials data facility (MDF) operates two cloudhosted services, data publication and data discovery, with features to promote open data sharing, self-service data publication and curation, and encourage data reuse, layered with powerful data discovery tools. The data publication service simplifies the process of copying data to a secure storage location, assigning data a citable persistent identifier, and recording custom (e.g., material, technique, or instrument specific)andmore » automatically-extractedmetadata in a registrywhile the data discovery service will provide advanced search capabilities (e.g., faceting, free text range querying, and full text search) against the registered data and metadata. TheMDF services empower individual researchers, research projects, and institutions to (I) publish research datasets, regardless of size, from local storage, institutional data stores, or cloud storage, without involvement of thirdparty publishers; (II) build, share, and enforce extensible domain-specific custom metadata schemas; (III) interact with published data and metadata via representational state transfer (REST) application program interfaces (APIs) to facilitate automation, analysis, and feedback; and (IV) access a data discovery model that allows researchers to search, interrogate, and eventually build on existing published data. We describe MDF’s design, current status, and future plans.« less

  7. The Materials Data Facility: Data Services to Advance Materials Science Research

    NASA Astrophysics Data System (ADS)

    Blaiszik, B.; Chard, K.; Pruyne, J.; Ananthakrishnan, R.; Tuecke, S.; Foster, I.

    2016-08-01

    With increasingly strict data management requirements from funding agencies and institutions, expanding focus on the challenges of research replicability, and growing data sizes and heterogeneity, new data needs are emerging in the materials community. The materials data facility (MDF) operates two cloud-hosted services, data publication and data discovery, with features to promote open data sharing, self-service data publication and curation, and encourage data reuse, layered with powerful data discovery tools. The data publication service simplifies the process of copying data to a secure storage location, assigning data a citable persistent identifier, and recording custom (e.g., material, technique, or instrument specific) and automatically-extracted metadata in a registry while the data discovery service will provide advanced search capabilities (e.g., faceting, free text range querying, and full text search) against the registered data and metadata. The MDF services empower individual researchers, research projects, and institutions to (I) publish research datasets, regardless of size, from local storage, institutional data stores, or cloud storage, without involvement of third-party publishers; (II) build, share, and enforce extensible domain-specific custom metadata schemas; (III) interact with published data and metadata via representational state transfer (REST) application program interfaces (APIs) to facilitate automation, analysis, and feedback; and (IV) access a data discovery model that allows researchers to search, interrogate, and eventually build on existing published data. We describe MDF's design, current status, and future plans.

  8. CyREST: Turbocharging Cytoscape Access for External Tools via a RESTful API.

    PubMed

    Ono, Keiichiro; Muetze, Tanja; Kolishovski, Georgi; Shannon, Paul; Demchak, Barry

    2015-01-01

    As bioinformatic workflows become increasingly complex and involve multiple specialized tools, so does the difficulty of reliably reproducing those workflows. Cytoscape is a critical workflow component for executing network visualization, analysis, and publishing tasks, but it can be operated only manually via a point-and-click user interface. Consequently, Cytoscape-oriented tasks are laborious and often error prone, especially with multistep protocols involving many networks. In this paper, we present the new cyREST Cytoscape app and accompanying harmonization libraries. Together, they improve workflow reproducibility and researcher productivity by enabling popular languages (e.g., Python and R, JavaScript, and C#) and tools (e.g., IPython/Jupyter Notebook and RStudio) to directly define and query networks, and perform network analysis, layouts and renderings. We describe cyREST's API and overall construction, and present Python- and R-based examples that illustrate how Cytoscape can be integrated into large scale data analysis pipelines. cyREST is available in the Cytoscape app store (http://apps.cytoscape.org) where it has been downloaded over 1900 times since its release in late 2014.

  9. CyREST: Turbocharging Cytoscape Access for External Tools via a RESTful API

    PubMed Central

    Ono, Keiichiro; Muetze, Tanja; Kolishovski, Georgi; Shannon, Paul; Demchak, Barry

    2015-01-01

    As bioinformatic workflows become increasingly complex and involve multiple specialized tools, so does the difficulty of reliably reproducing those workflows. Cytoscape is a critical workflow component for executing network visualization, analysis, and publishing tasks, but it can be operated only manually via a point-and-click user interface. Consequently, Cytoscape-oriented tasks are laborious and often error prone, especially with multistep protocols involving many networks. In this paper, we present the new cyREST Cytoscape app and accompanying harmonization libraries. Together, they improve workflow reproducibility and researcher productivity by enabling popular languages (e.g., Python and R, JavaScript, and C#) and tools (e.g., IPython/Jupyter Notebook and RStudio) to directly define and query networks, and perform network analysis, layouts and renderings. We describe cyREST’s API and overall construction, and present Python- and R-based examples that illustrate how Cytoscape can be integrated into large scale data analysis pipelines. cyREST is available in the Cytoscape app store (http://apps.cytoscape.org) where it has been downloaded over 1900 times since its release in late 2014. PMID:26672762

  10. High-Throughput and Low-Latency Network Communication with NetIO

    NASA Astrophysics Data System (ADS)

    Schumacher, Jörn; Plessl, Christian; Vandelli, Wainer

    2017-10-01

    HPC network technologies like Infiniband, TrueScale or OmniPath provide low- latency and high-throughput communication between hosts, which makes them attractive options for data-acquisition systems in large-scale high-energy physics experiments. Like HPC networks, DAQ networks are local and include a well specified number of systems. Unfortunately traditional network communication APIs for HPC clusters like MPI or PGAS exclusively target the HPC community and are not suited well for DAQ applications. It is possible to build distributed DAQ applications using low-level system APIs like Infiniband Verbs, but it requires a non-negligible effort and expert knowledge. At the same time, message services like ZeroMQ have gained popularity in the HEP community. They make it possible to build distributed applications with a high-level approach and provide good performance. Unfortunately, their usage usually limits developers to TCP/IP- based networks. While it is possible to operate a TCP/IP stack on top of Infiniband and OmniPath, this approach may not be very efficient compared to a direct use of native APIs. NetIO is a simple, novel asynchronous message service that can operate on Ethernet, Infiniband and similar network fabrics. In this paper the design and implementation of NetIO is presented and described, and its use is evaluated in comparison to other approaches. NetIO supports different high-level programming models and typical workloads of HEP applications. The ATLAS FELIX project [1] successfully uses NetIO as its central communication platform. The architecture of NetIO is described in this paper, including the user-level API and the internal data-flow design. The paper includes a performance evaluation of NetIO including throughput and latency measurements. The performance is compared against the state-of-the- art ZeroMQ message service. Performance measurements are performed in a lab environment with Ethernet and FDR Infiniband networks.

  11. When Will It Be …?: U.S. Naval Observatory Calendar Computers

    NASA Astrophysics Data System (ADS)

    Bartlett, Jennifer L.; Chizek Frouard, Malynda; Lesniak, Michael V.

    2016-06-01

    Sensitivity to religious calendars is increasingly expected when planning activities. Consequently, the U.S. Naval Observatory (USNO) has redesigned its on-line calendar resources to allow the computation of select religious dates for specific years via an application programming interface (API). This flexible interface returns dates in JavaScript Object Notation (JSON) that can be incorporated into third-party websites or applications. Currently, the services compute Christian, Islamic, and Jewish events.The “Dates of Ash Wednesday and Easter” service (http://aa.usno.navy.mil/data/docs/easter.php) returns the dates of these two events for years after 1582 C.E. (1582 A.D.) The method of the western Christian churches is used to determined when Easter, a moveable feast, occurs.The “Dates of Islamic New Year and Ramadan” service (http://aa.usno.navy.mil/data/docs/islamic.php) returns the approximate Gregorian dates of these two events for years after 1582 C.E. (990 A.H.) and Julian dates are computed for the years 622-1582 C.E. (1-990 A.H.). The appropriate year in the Islamic calendar (anno Hegira) is also provided. Each event begins at 6 P.M. or sunset on the preceding day. These events are computed using a tabular calendar for planning purposes; in practice, the actual event is determined by observation of the appropriate new Moon.The “First Day of Passover” service (http://aa.usno.navy.mil/data/docs/passover.php) returns the Gregorian date corresponding to Nisan 15 for years after 1582 C.E. (5342 A.M.) and Julian dates are computed for the years 360-1582 C.E. (4120-5342 A.M.). The appropriate year in the Jewish calendar (anno Mundi) is also provided. Passover begins at 6 P.M. or sunset on the preceding day.On-line documentation for using the API-enabled calendar computers, including sample calls, is available (http://aa.usno.navy.mil/data/docs/api.php). The same web page also describes how to reach the Complete Sun and Moon Data for One Day, Phases of the Moon, Solar Eclipse Computer, Day and Night Across the Earth, and Apparent Disk of a Solar System Object services using API calls.An “Introduction to Calendars” (http://aa.usno.navy.mil/faq/docs/calendars.php) provides an overview of the topic and links to additional resources.

  12. A JEE RESTful service to access Conditions Data in ATLAS

    NASA Astrophysics Data System (ADS)

    Formica, Andrea; Gallas, E. J.

    2015-12-01

    Usage of condition data in ATLAS is extensive for offline reconstruction and analysis (e.g. alignment, calibration, data quality). The system is based on the LCG Conditions Database infrastructure, with read and write access via an ad hoc C++ API (COOL), a system which was developed before Run 1 data taking began. The infrastructure dictates that the data is organized into separate schemas (assigned to subsystems/groups storing distinct and independent sets of conditions), making it difficult to access information from several schemas at the same time. We have thus created PL/SQL functions containing queries to provide content extraction at multi-schema level. The PL/SQL API has been exposed to external clients by means of a Java application providing DB access via REST services, deployed inside an application server (JBoss WildFly). The services allow navigation over multiple schemas via simple URLs. The data can be retrieved either in XML or JSON formats, via simple clients (like curl or Web browsers).

  13. Aedes communis Reactivity Is Associated with Bee Venom Hypersensitivity: An in vitro and in vivo Study.

    PubMed

    Scala, Enrico; Pirrotta, Lia; Uasuf, Carina G; Mistrello, Gianni; Amato, Stefano; Guerra, Emma Cristina; Locanto, Maria; Meneguzzi, Giorgia; Giani, Mauro; Cecchi, Lorenzo; Abeni, Damiano; Asero, Riccardo

    2018-01-01

    Mosquito bite is usually followed by a local reaction, but severe or systemic reaction may, in rare cases, occur. Allergic reactions to Aedes communis (Ac) may be underestimated due to the lack of reliable diagnostic tools. In this multicenter study, 205 individuals reporting large local reactions to Ac were enrolled and studied for cutaneous or IgE reactivity to Ac, Blattella germanica, Penaeus monodon, and Dermatophagoides pteronyssinus. Extract and molecular IgE reactivity to bees, wasps, hornets, and yellow jacket venoms were also studied in 119 patients with a clinical history of adverse reaction to Hymenoptera. Immunoblot (IB) analysis and immunoCAP IgE inhibition experiments were carried out in selected sera. Ac sensitization was recorded in 96 (46.8%) patients on SPT. Strict relationship between Ac and D. pteronyssinus, B. germanica, P. monodon, or Apis mellifera reactivity on SPT was observed. Ac IgE recognition was seen in 60/131 (45.8%) patients, 49 (81.6%) of them SPT positive, and 5/14 IB reactors. Ac IgE sensitization was associated with Tabanus spp, A. mellifera, Vespula vulgaris, and Polistes dominula reactivity. A strict relationship between Ac IgE reactivity and Api m 1, Api m 2, Api m 3, Api m 5, and Api m 10 was recorded. IgE reactivity to AC was inhibited in 9/15 cases after serum absorption with the A. mellifera extract. Both SPT and IgE Ac reactivity is observed in about half of patients with a history of large local reactions to mosquito bites. The significant relationship between Ac sensitization and either extract or single bee venom components is suggestive of a "bee-mosquito syndrome" occurrence. © 2018 S. Karger AG, Basel.

  14. The User Community and a Multi-Mission Data Project: Services, Experiences and Directions of the Space Physics Data Facility

    NASA Technical Reports Server (NTRS)

    Fung, Shing F.; Bilitza, D.; Candey, R.; Chimiak, R.; Cooper, John; Fung, Shing; Harris, B.; Johnson R.; King, J.; Kovalick, T.; hide

    2008-01-01

    From a user's perspective, the multi-mission data and orbit services of NASA's Space Physics Data Facility (SPDF) project offer a unique range of important data and services highly complementary to other services presently available or now evolving in the international heliophysics data environment. The VSP (Virtual Space Physics Observatory) service is an active portal to a wide range of distributed data sources. CDAWeb (Coordinate Data Analysis Web) enables plots, listings and file downloads for current data cross the boundaries of missions and instrument types (and now including data from THEMIS and STEREO). SSCWeb, Helioweb and our 3D Animated Orbit Viewer (TIPSOD) provide position data and query logic for most missions currently important to heliophysics science. OMNIWeb with its new extension to 1- and 5-minute resolution provides interplanetary parameters at the Earth's bow shock as a unique value-added data product. SPDF also maintains NASA's CDF (common Data Format) standard and a range of associated tools including translation services. These capabilities are all now available through webservices-based APIs as well as through our direct user interfaces. In this paper, we will demonstrate the latest data and capabilities now supported in these multi-mission services, review the lessons we continue to learn in what science users need and value in this class of services, and discuss out current thinking to the future role and appropriate focus of the SPDF effort in the evolving and increasingly distributed heliophysics data environment.

  15. Search without Boundaries Using Simple APIs

    USGS Publications Warehouse

    Tong, Qi

    2009-01-01

    The U.S. Geological Survey (USGS) Library, where the author serves as the digital services librarian, is increasingly challenged to make it easier for users to find information from many heterogeneous information sources. Information is scattered throughout different software applications (i.e., library catalog, federated search engine, link resolver, and vendor websites), and each specializes in one thing. How could the library integrate the functionalities of one application with another and provide a single point of entry for users to search across? To improve the user experience, the library launched an effort to integrate the federated search engine into the library's intranet website. The result is a simple search box that leverages the federated search engine's built-in application programming interfaces (APIs). In this article, the author describes how this project demonstrated the power of APIs and their potential to be used by other enterprise search portals inside or outside of the library.

  16. Ingredientes Farmacéuticos Activos Potencialmente Inapropiados en Adultos Mayores: Lista IFAsPIAM: Panel de Consenso Argentino.

    PubMed

    Marzi, Marta M; Pires, Miryam S; Quaglia, Nora B

    2018-04-18

    To perform a list agreed by Argentinean experts and adapted to the local context containing potentially inappropriate (PI) medications in old people (OP) usingthe Delphi consensus technique optimized for this subject. A preliminary list of potentially inappropriate medications (PIM) was drawn up based on foreign PIM lists and a selective search in the scientific literature. The iterative Delphi process was used to submit the active pharmaceutical ingredients (APIs) of the preliminary PIM list to the panel of Argentinean experts. The analysis of theanswers to determine the arrival to the consensus was carried out applying three criteria specially defined for this purpose. After two Delphi rounds, it was not reached agreement about 12 APIs. The List of explicit criteria for PIAPIs for use in OP (IFAsPIAM List) was finally constituted by 128 APIs corresponding to 9 groups of the ATC classification system to which they were organized. In addition to each API, information justifying the unfavorable benefit/risk profile and therapeutic alternatives or recommendations/precautions was recorded. The group with the most PI APIs was N (NervousSystem) (60; 47%) followed by groups C (Cardiovascular) and M (Musculoskeletal). This study presents the first Latin American list of PIM in OP developed using an expert consensus technique. The IFAs PIAM List would contribute to the rational use of drugs in elderly population, constituting a valuable tool in Argentinean public health. Copyright © 2018. Published by Elsevier Inc.

  17. Physical stability of API/polymer-blend amorphous solid dispersions.

    PubMed

    Lehmkemper, Kristin; Kyeremateng, Samuel O; Bartels, Mareike; Degenhardt, Matthias; Sadowski, Gabriele

    2018-03-01

    The preparation of amorphous solid dispersions (ASDs) is a well-established strategy for formulating active pharmaceutical ingredients by embedding them in excipients, usually amorphous polymers. Different polymers can be combined for designing ASDs with desired properties like an optimized dissolution behavior. One important criterion for the development of ASD compositions is the physical stability. In this work, the physical stability of API/polymer-blend ASDs was investigated by thermodynamic modeling and stability studies. Amorphous naproxen (NAP) and acetaminophen (APAP) were embedded in blends of hydroxypropyl methylcellulose acetate succinate (HPMCAS) and either poly(vinylpyrrolidone) (PVP) or poly(vinylpyrrolidone-co-vinyl acetate) (PVPVA64). Parameters for modeling the API solubility in the blends and the glass-transition temperature curves of the water-free systems with Perturbed-Chain Statistical Associating Fluid Theory and Kwei equation, respectively, were correlated to experimental data. The phase behavior for standardized storage conditions (0%, 60% and 75% relative humidity (RH)) was predicted and compared to six months-long stability studies. According to modeling and experimental results, the physical stability was reduced with increasing HPMCAS content and increasing RH. This trend was observed for all investigated systems, with both APIs (NAP and APAP) and both polymer blends (PVP/HPMCAS and PVPVA64/HPMCAS). PC-SAFT and the Kwei equation turned out to be suitable tools for modeling and predicting the physical stability of the investigated API/polymer-blends ASDs. Copyright © 2017 Elsevier B.V. All rights reserved.

  18. Isolation of Mal d 1 and Api g 1 - specific recombinant antibodies from mouse IgG Fab fragment libraries - Mal d 1-specific antibody exhibits cross-reactivity against Bet v 1.

    PubMed

    Haka, Jaana; Niemi, Merja H; Iljin, Kristiina; Reddy, Vanga Siva; Takkinen, Kristiina; Laukkanen, Marja-Leena

    2015-05-27

    Around 3-5% of the population suffer from IgE-mediated food allergies in Western countries and the number of food-allergenic people is increasing. Individuals with certain pollen allergies may also suffer from a sensitisation to proteins in the food products. As an example a person sensitised to the major birch pollen allergen, Bet v 1, is often sensitised to its homologues, such as the major allergens of apple, Mal d 1, and celery, Api g 1, as well. Development of tools for the reliable, sensitive and quick detection of allergens present in various food products is essential for allergic persons to prevent the consumption of substances causing mild and even life-threatening immune responses. The use of monoclonal antibodies would ensure the specific detection of the harmful food content for a sensitised person. Mouse IgG antibody libraries were constructed from immunised mice and specific recombinant antibodies for Mal d 1 and Api g 1 were isolated from the libraries by phage display. More detailed characterisation of the resulting antibodies was carried out using ELISA, SPR experiments and immunoprecipitation assays. The allergen-specific Fab fragments exhibited high affinity towards the target recombinant allergens. Furthermore, the Fab fragments also recognised native allergens from natural sources. Interestingly, isolated Mal d 1-specific antibody bound also to Bet v 1, the main allergen eliciting the cross-reactivity syndrome between the birch pollen and apple. Despite the similarities in Api g 1 and Bet v 1 tertiary structures, the isolated Api g 1-specific antibodies showed no cross-reactivity to Bet v 1. Here, high-affinity allergen-specific recombinant antibodies were isolated with interesting binding properties. With further development, these antibodies can be utilised as tools for the specific and reliable detection of allergens from different consumable products. This study gives new preliminary insights to elucidate the mechanism behind the pollen-food syndrome and to study the IgG epitope of the allergens.

  19. Detection of Apoptosis in Early Life Stages as a Tool to Evaluate Chemical Control of Invasive Species

    DTIC Science & Technology

    2007-08-01

    ERDC/TN ANSRP-07-2 August 2007 Detection of Apoptosis in Early Life Stages as a Tool to Evaluate Chemical Control of Invasive Species by J...4. TITLE AND SUBTITLE Detection of Apoptosis in Early Life Stages as a Tool to Evaluate Chemical Control of Invasive Species 5a. CONTRACT NUMBER 5b...heralding apoptosis . Data analysis. An apoptotic index (API) was established by calculating the percentage of embryos in each life stage with

  20. Design and Implementation of KSP on the Next Generation Cryptography API

    NASA Astrophysics Data System (ADS)

    Lina, Zhang

    With good seamless connectivity and higher safety, KSP (Key Storage Providers) is the inexorable trend of security requirements and development to take the place of CSP (Cryptographic Service Provider). But the study on KSP has just started in our country, and almost no reports of its implementation can be found. Based on the analysis of function modules and the architecture of Cryptography API (Next Generation (CNG)), this paper discusses the design and implementation of KSP (key storage providers) based on smart card in detail, and an example is also presented to illustrate how to use KSP in Windows Vista.

  1. 49 CFR 195.205 - Repair, alteration and reconstruction of aboveground breakout tanks that have been in service.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ..., examination, and material requirements of those respective standards. (3) For high pressure tanks built to API... service must be capable of withstanding the internal pressure produced by the hazardous liquid to be... approximately atmospheric pressure constructed of carbon and low alloy steel, welded or riveted, and non...

  2. Geneious Basic: An integrated and extendable desktop software platform for the organization and analysis of sequence data

    PubMed Central

    Kearse, Matthew; Moir, Richard; Wilson, Amy; Stones-Havas, Steven; Cheung, Matthew; Sturrock, Shane; Buxton, Simon; Cooper, Alex; Markowitz, Sidney; Duran, Chris; Thierer, Tobias; Ashton, Bruce; Meintjes, Peter; Drummond, Alexei

    2012-01-01

    Summary: The two main functions of bioinformatics are the organization and analysis of biological data using computational resources. Geneious Basic has been designed to be an easy-to-use and flexible desktop software application framework for the organization and analysis of biological data, with a focus on molecular sequences and related data types. It integrates numerous industry-standard discovery analysis tools, with interactive visualizations to generate publication-ready images. One key contribution to researchers in the life sciences is the Geneious public application programming interface (API) that affords the ability to leverage the existing framework of the Geneious Basic software platform for virtually unlimited extension and customization. The result is an increase in the speed and quality of development of computation tools for the life sciences, due to the functionality and graphical user interface available to the developer through the public API. Geneious Basic represents an ideal platform for the bioinformatics community to leverage existing components and to integrate their own specific requirements for the discovery, analysis and visualization of biological data. Availability and implementation: Binaries and public API freely available for download at http://www.geneious.com/basic, implemented in Java and supported on Linux, Apple OSX and MS Windows. The software is also available from the Bio-Linux package repository at http://nebc.nerc.ac.uk/news/geneiousonbl. Contact: peter@biomatters.com PMID:22543367

  3. Geneious Basic: an integrated and extendable desktop software platform for the organization and analysis of sequence data.

    PubMed

    Kearse, Matthew; Moir, Richard; Wilson, Amy; Stones-Havas, Steven; Cheung, Matthew; Sturrock, Shane; Buxton, Simon; Cooper, Alex; Markowitz, Sidney; Duran, Chris; Thierer, Tobias; Ashton, Bruce; Meintjes, Peter; Drummond, Alexei

    2012-06-15

    The two main functions of bioinformatics are the organization and analysis of biological data using computational resources. Geneious Basic has been designed to be an easy-to-use and flexible desktop software application framework for the organization and analysis of biological data, with a focus on molecular sequences and related data types. It integrates numerous industry-standard discovery analysis tools, with interactive visualizations to generate publication-ready images. One key contribution to researchers in the life sciences is the Geneious public application programming interface (API) that affords the ability to leverage the existing framework of the Geneious Basic software platform for virtually unlimited extension and customization. The result is an increase in the speed and quality of development of computation tools for the life sciences, due to the functionality and graphical user interface available to the developer through the public API. Geneious Basic represents an ideal platform for the bioinformatics community to leverage existing components and to integrate their own specific requirements for the discovery, analysis and visualization of biological data. Binaries and public API freely available for download at http://www.geneious.com/basic, implemented in Java and supported on Linux, Apple OSX and MS Windows. The software is also available from the Bio-Linux package repository at http://nebc.nerc.ac.uk/news/geneiousonbl.

  4. Software Applications to Access Earth Science Data: Building an ECHO Client

    NASA Astrophysics Data System (ADS)

    Cohen, A.; Cechini, M.; Pilone, D.

    2010-12-01

    Historically, developing an ECHO (NASA’s Earth Observing System (EOS) ClearingHOuse) client required interaction with its SOAP API. SOAP, as a framework for web service communication has numerous advantages for Enterprise applications and Java/C# type programming languages. However, as interest has grown for quick development cycles and more intriguing “mashups,” ECHO has seen the SOAP API lose its appeal. In order to address these changing needs, ECHO has introduced two new interfaces facilitating simple access to its metadata holdings. The first interface is built upon the OpenSearch format and ESIP Federated Search framework. The second interface is built upon the Representational State Transfer (REST) architecture. Using the REST and OpenSearch APIs to access ECHO makes development with modern languages much more feasible and simpler. Client developers can leverage the simple interaction with ECHO to focus more of their time on the advanced functionality they are presenting to users. To demonstrate the simplicity of developing with the REST API, participants will be led through a hands-on experience where they will develop an ECHO client that performs the following actions: + Login + Provider discovery + Provider based dataset discovery + Dataset, Temporal, and Spatial constraint based Granule discovery + Online Data Access

  5. Design and Implement AN Interoperable Internet of Things Application Based on AN Extended Ogc Sensorthings Api Standard

    NASA Astrophysics Data System (ADS)

    Huang, C. Y.; Wu, C. H.

    2016-06-01

    The Internet of Things (IoT) is an infrastructure that interconnects uniquely-identifiable devices using the Internet. By interconnecting everyday appliances, various monitoring and physical mashup applications can be constructed to improve people's daily life. However, IoT devices created by different manufacturers follow different proprietary protocols and cannot communicate with each other. This heterogeneity issue causes different products to be locked in multiple closed ecosystems that we call IoT silos. In order to address this issue, a common industrial solution is the hub approach, which implements connectors to communicate with IoT devices following different protocols. However, with the growing number of proprietary protocols proposed by device manufacturers, IoT hubs need to support and maintain a lot of customized connectors. Hence, we believe the ultimate solution to address the heterogeneity issue is to follow open and interoperable standard. Among the existing IoT standards, the Open Geospatial Consortium (OGC) SensorThings API standard supports comprehensive conceptual model and query functionalities. The first version of SensorThings API mainly focuses on connecting to IoT devices and sharing sensor observations online, which is the sensing capability. Besides the sensing capability, IoT devices could also be controlled via the Internet, which is the tasking capability. While the tasking capability was not included in the first version of the SensorThings API standard, this research aims on defining the tasking capability profile and integrates with the SensorThings API standard, which we call the extended-SensorThings API in this paper. In general, this research proposes a lightweight JSON-based web service description, the "Tasking Capability Description", allowing device owners and manufacturers to describe different IoT device protocols. Through the extended- SensorThings API, users and applications can follow a coherent protocol to control IoT devices that use different communication protocols, which could consequently achieve the interoperable Internet of Things infrastructure.

  6. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jarocki, John Charles; Zage, David John; Fisher, Andrew N.

    LinkShop is a software tool for applying the method of Linkography to the analysis time-sequence data. LinkShop provides command line, web, and application programming interfaces (API) for input and processing of time-sequence data, abstraction models, and ontologies. The software creates graph representations of the abstraction model, ontology, and derived linkograph. Finally, the tool allows the user to perform statistical measurements of the linkograph and refine the ontology through direct manipulation of the linkograph.

  7. APIS : an interactive database of HST-UV observations of the outer planets

    NASA Astrophysics Data System (ADS)

    Lamy, Laurent; Henry, Florence; Prangé, Renée; Le Sidaner, Pierre

    2014-05-01

    Remote UV measurement of the outer planets offer a wealth of informations on rings, moons, planetary atmospheres and magnetospheres. Auroral emissions in particular provide highly valuable constraints on the auroral processes at work and the underlying coupling between the solar wind, the magnetosphere, the ionosphere and the moons. Key observables provided by high resolution spectro-imaging include the spatial topology and the dynamics of active magnetic field lines, the radiated and the precipitated powers or the energy of precipitating particles. The Hubble Space Telescope (HST) acquired thousands of Far-UV spectra and images of the aurorae of Jupiter, Saturn and Uranus since 1993, feeding in numerous magnetospheric studies. But their use remains generally limited, owing to the difficulty to access and use raw and value-added data. APIS, the egyptian god of fertilization, is also the acronym of a new database (Auroral Planetary Imaging and Spectroscopy), aimed at facilitating the use of HST planetary auroral observations. APIS is based at the Virtual Observatory (VO) of Paris and provides a free and interactive access to a variety of high level data through a simple research interface and standard VO tools (as Aladin, Specview). We will present the capabilities of APIS and illustrate them with several examples.

  8. APIS : an interactive database of HST-UV observations of the outer planets

    NASA Astrophysics Data System (ADS)

    Lamy, L.; Henry, F.; Prangé, R.; Le Sidaner, P.

    2013-09-01

    Remote UV measurement of the outer planets are a wealth of informations on rings, moons, planetary atmospheres and magnetospheres. Auroral emissions in particular provide highly valuable constraints on the auroral processes at work and the underlying coupling between the solar wind, the magnetosphere, the ionosphere and the moons. Key observables provided by high resolution spectro-imaging include the spatial topology and the dynamics of active magnetic field lines, the radiated and the precipitated powers or the energy of precipitating particles. The Hubble Space Telescope (HST) acquired thousands of Far-UV spectra and images of the aurorae of Jupiter, Saturn and Uranus since 1993, feeding in numerous magnetospheric studies. But their use remains generally limited, owing to the difficulty to access and use raw and value-added data. APIS, the egyptian god of fertilization, is also the acronym of a new database (Auroral Planetary Imaging and Spectroscopy, Figure 1), aimed at facilitating the use of HST planetary auroral observations. APIS is based at the Virtual Observatory (VO) of Paris and provides a free and interactive access to a variety of high level data through a simple research interface and standard VO tools. We will present the capabilities of APIS and illustrate them with several examples.

  9. Early pharmaceutical profiling to predict oral drug absorption: current status and unmet needs.

    PubMed

    Bergström, Christel A S; Holm, René; Jørgensen, Søren Astrup; Andersson, Sara B E; Artursson, Per; Beato, Stefania; Borde, Anders; Box, Karl; Brewster, Marcus; Dressman, Jennifer; Feng, Kung-I; Halbert, Gavin; Kostewicz, Edmund; McAllister, Mark; Muenster, Uwe; Thinnes, Julian; Taylor, Robert; Mullertz, Anette

    2014-06-16

    Preformulation measurements are used to estimate the fraction absorbed in vivo for orally administered compounds and thereby allow an early evaluation of the need for enabling formulations. As part of the Oral Biopharmaceutical Tools (OrBiTo) project, this review provides a summary of the pharmaceutical profiling methods available, with focus on in silico and in vitro models typically used to forecast active pharmaceutical ingredient's (APIs) in vivo performance after oral administration. An overview of the composition of human, animal and simulated gastrointestinal (GI) fluids is provided and state-of-the art methodologies to study API properties impacting on oral absorption are reviewed. Assays performed during early development, i.e. physicochemical characterization, dissolution profiles under physiological conditions, permeability assays and the impact of excipients on these properties are discussed in detail and future demands on pharmaceutical profiling are identified. It is expected that innovative computational and experimental methods that better describe molecular processes involved in vivo during dissolution and absorption of APIs will be developed in the OrBiTo. These methods will provide early insights into successful pathways (medicinal chemistry or formulation strategy) and are anticipated to increase the number of new APIs with good oral absorption being discovered. Copyright © 2013 Elsevier B.V. All rights reserved.

  10. Ultrasound-assisted powder-coating technique to improve content uniformity of low-dose solid dosage forms.

    PubMed

    Genina, Natalja; Räikkönen, Heikki; Antikainen, Osmo; Heinämäki, Jyrki; Yliruusi, Jouko

    2010-09-01

    An ultrasound-assisted powder-coating technique was used to produce a homogeneous powder formulation of a low-dose active pharmaceutical ingredient (API). The powdered particles of microcrystalline cellulose (MCC; Avicel® PH-200) were coated with a 4% m/V aqueous solution of riboflavin sodium phosphate, producing a uniform drug layer on the particle surfaces. It was possible to regulate the amount of API in the treated powder. The thickness of the API layer on the surface of the MCC particles increased near linearly as the number of coating cycles increased, allowing a precise control of the drug content. The tablets (n = 950) prepared from the coated powder showed significantly improved weight and content uniformity in comparison with the reference tablets compressed from a physical binary powder mixture. This was due to the coated formulation remaining uniform during the entire tabletting process, whereas the physical mixture of the powders was subject to segregation. In conclusion, the ultrasound-assisted technique presented here is an effective tool for homogeneous drug coating of powders of irregular particle shape and broad particle size distribution, improving content uniformity of low-dose API in tablets, and consequently, ensuring the safe delivery of a potent active substance to patients.

  11. Genetics Home Reference: Pallister-Killian mosaic syndrome

    MedlinePlus

    ... qualified healthcare professional . About Selection Criteria for Links Data Files & API Site Map Subscribe Customer Support USA.gov Copyright Privacy Accessibility FOIA Viewers & Players U.S. Department of Health & Human Services National Institutes of Health National Library of ...

  12. Extending Climate Analytics-As to the Earth System Grid Federation

    NASA Astrophysics Data System (ADS)

    Tamkin, G.; Schnase, J. L.; Duffy, D.; McInerney, M.; Nadeau, D.; Li, J.; Strong, S.; Thompson, J. H.

    2015-12-01

    We are building three extensions to prior-funded work on climate analytics-as-a-service that will benefit the Earth System Grid Federation (ESGF) as it addresses the Big Data challenges of future climate research: (1) We are creating a cloud-based, high-performance Virtual Real-Time Analytics Testbed supporting a select set of climate variables from six major reanalysis data sets. This near real-time capability will enable advanced technologies like the Cloudera Impala-based Structured Query Language (SQL) query capabilities and Hadoop-based MapReduce analytics over native NetCDF files while providing a platform for community experimentation with emerging analytic technologies. (2) We are building a full-featured Reanalysis Ensemble Service comprising monthly means data from six reanalysis data sets. The service will provide a basic set of commonly used operations over the reanalysis collections. The operations will be made accessible through NASA's climate data analytics Web services and our client-side Climate Data Services (CDS) API. (3) We are establishing an Open Geospatial Consortium (OGC) WPS-compliant Web service interface to our climate data analytics service that will enable greater interoperability with next-generation ESGF capabilities. The CDS API will be extended to accommodate the new WPS Web service endpoints as well as ESGF's Web service endpoints. These activities address some of the most important technical challenges for server-side analytics and support the research community's requirements for improved interoperability and improved access to reanalysis data.

  13. PChopper: high throughput peptide prediction for MRM/SRM transition design.

    PubMed

    Afzal, Vackar; Huang, Jeffrey T-J; Atrih, Abdel; Crowther, Daniel J

    2011-08-15

    The use of selective reaction monitoring (SRM) based LC-MS/MS analysis for the quantification of phosphorylation stoichiometry has been rapidly increasing. At the same time, the number of sites that can be monitored in a single LC-MS/MS experiment is also increasing. The manual processes associated with running these experiments have highlighted the need for computational assistance to quickly design MRM/SRM candidates. PChopper has been developed to predict peptides that can be produced via enzymatic protein digest; this includes single enzyme digests, and combinations of enzymes. It also allows digests to be simulated in 'batch' mode and can combine information from these simulated digests to suggest the most appropriate enzyme(s) to use. PChopper also allows users to define the characteristic of their target peptides, and can automatically identify phosphorylation sites that may be of interest. Two application end points are available for interacting with the system; the first is a web based graphical tool, and the second is an API endpoint based on HTTP REST. Service oriented architecture was used to rapidly develop a system that can consume and expose several services. A graphical tool was built to provide an easy to follow workflow that allows scientists to quickly and easily identify the enzymes required to produce multiple peptides in parallel via enzymatic digests in a high throughput manner.

  14. Fast Multiscale Algorithms for Information Representation and Fusion

    DTIC Science & Technology

    2011-07-01

    We are also developing convenient command-line invocation tools in addition to the previously developed APIs . Various real-world data sets...This knowledge is important in geolocation applications where knowing whether a received signal is line-of-sight or not is necessary for the

  15. Hadoop-BAM: directly manipulating next generation sequencing data in the cloud

    PubMed Central

    Niemenmaa, Matti; Kallio, Aleksi; Schumacher, André; Klemelä, Petri; Korpelainen, Eija; Heljanko, Keijo

    2012-01-01

    Summary: Hadoop-BAM is a novel library for the scalable manipulation of aligned next-generation sequencing data in the Hadoop distributed computing framework. It acts as an integration layer between analysis applications and BAM files that are processed using Hadoop. Hadoop-BAM solves the issues related to BAM data access by presenting a convenient API for implementing map and reduce functions that can directly operate on BAM records. It builds on top of the Picard SAM JDK, so tools that rely on the Picard API are expected to be easily convertible to support large-scale distributed processing. In this article we demonstrate the use of Hadoop-BAM by building a coverage summarizing tool for the Chipster genome browser. Our results show that Hadoop offers good scalability, and one should avoid moving data in and out of Hadoop between analysis steps. Availability: Available under the open-source MIT license at http://sourceforge.net/projects/hadoop-bam/ Contact: matti.niemenmaa@aalto.fi Supplementary information: Supplementary material is available at Bioinformatics online. PMID:22302568

  16. Runtime Performance Monitoring Tool for RTEMS System Software

    NASA Astrophysics Data System (ADS)

    Cho, B.; Kim, S.; Park, H.; Kim, H.; Choi, J.; Chae, D.; Lee, J.

    2007-08-01

    RTEMS is a commercial-grade real-time operating system that supports multi-processor computers. However, there are not many development tools for RTEMS. In this paper, we report new RTEMS-based runtime performance monitoring tool. We have implemented a light weight runtime monitoring task with an extension to the RTEMS APIs. Using our tool, software developers can verify various performance- related parameters during runtime. Our tool can be used during software development phase and in-orbit operation as well. Our implemented target agent is light weight and has small overhead using SpaceWire interface. Efforts to reduce overhead and to add other monitoring parameters are currently under research.

  17. Tailoring Earth Observation To Ranchers For Improved Land Management And Profitability: The VegMachine Online Project

    NASA Astrophysics Data System (ADS)

    Scarth, P.; Trevithick, B.; Beutel, T.

    2016-12-01

    VegMachine Online is a freely available browser application that allows ranchers across Australia to view and interact with satellite derived ground cover state and change maps on their property and extract this information in a graphical format using interactive tools. It supports the delivery and communication of a massive earth observation data set in an accessible, producer friendly way . Around 250,000 Landsat TM, ETM and OLI images were acquired across Australia, converted to terrain corrected surface reflectance and masked for cloud, cloud shadow, terrain shadow and water. More than 2500 field sites across the Australian rangelands were used to derive endmembers used in a constrained unmixing approach to estimate the per-pixel proportion of bare, green and non-green vegetation for all images. A seasonal metoid compositing method was used to produce national fractional cover virtual mosaics for each three month period since 1988. The time series of green fraction is used to estimate the persistent green due to tree and shrub canopies, and this estimate is used to correct the fractional cover to ground cover for our mixed tree-grass rangeland systems. Finally, deciles are produced for key metrics every season to track a pixels relativity to the entire time series. These data are delivered through time series enabled web mapping services and customised web processing services that enable the full time series over any spatial extent to be interrogated in seconds via a RESTful interface. These services interface with a front end browser application that provides product visualization for any date in the time series, tools to draw or import polygon boundaries, plot time series ground cover comparisons, look at the effect of historical rainfall and tools to run the revised universal soil loss equation in web time to assess the effect of proposed changes in cover retention. VegMachine Online is already being used by ranchers monitoring paddock condition, organisations supporting land management initiatives in Great Barrier Reef catchments, by students developing tools to understand land condition and degradation and the underlying data and APIs are supporting several other land condition mapping tools.

  18. Genetics Home Reference: dilated cardiomyopathy with ataxia syndrome

    MedlinePlus

    ... qualified healthcare professional . About Selection Criteria for Links Data Files & API Site Map Subscribe Customer Support USA.gov Copyright Privacy Accessibility FOIA Viewers & Players U.S. Department of Health & Human Services National Institutes of Health National Library of ...

  19. The Demonstrator for the European Plate Observing System (EPOS)

    NASA Astrophysics Data System (ADS)

    Hoffmann, T. L.; Euteneuer, F.; Ulbricht, D.; Lauterjung, J.; Bailo, D.; Jeffery, K. G.

    2014-12-01

    An important outcome of the 4-year Preparatory Phase of the ESFRI project European Plate Observing System (EPOS) was the development and first implementation of the EPOS Demonstrator by the project's ICT Working Group 7. The Demonstrator implements the vertical integration of the three-layer architectural scheme for EPOS, connecting the Integrated Core Services (ICS), Thematic Core Services (TCS) and the National Research Infrastructures (NRI). The demonstrator provides a single GUI with central key discovery and query functionalities, based on already existing services by the seismic, geologic and geodetic communities. More specifically the seismic services of the Demonstrator utilize webservices and APIs for data and discovery of raw seismic data (FDSN webservices by the EIDA Network), events (Geoportal by EMSC) and analytical data products (e.g., hazard maps by EFEHR via OGC WMS). For geologic services, the EPOS Demonstrator accesses OneGeology Europe which serves the community with geologic maps and point information via OGC webservices. The Demonstrator also provides access to raw geodetic data via a newly developed universal tool called GSAC. The Demonstrator itself resembles the future Integrated Core Service (ICS) and provides direct access to the end user. Its core functionality lies in a metadata catalogue, which serves as the central information hub and stores information about all RIs, related persons, projects, financial background and technical access information. The database schema of the catalogue is based on CERIF, which has been slightly adapted. Currently, the portal provides basic query functions as well as cross domain search. [www.epos.cineca.it

  20. Development of the geometry database for the CBM experiment

    NASA Astrophysics Data System (ADS)

    Akishina, E. P.; Alexandrov, E. I.; Alexandrov, I. N.; Filozova, I. A.; Friese, V.; Ivanov, V. V.

    2018-01-01

    The paper describes the current state of the Geometry Database (Geometry DB) for the CBM experiment. The main purpose of this database is to provide convenient tools for: (1) managing the geometry modules; (2) assembling various versions of the CBM setup as a combination of geometry modules and additional files. The CBM users of the Geometry DB may use both GUI (Graphical User Interface) and API (Application Programming Interface) tools for working with it.

  1. AnnoLnc: a web server for systematically annotating novel human lncRNAs.

    PubMed

    Hou, Mei; Tang, Xing; Tian, Feng; Shi, Fangyuan; Liu, Fenglin; Gao, Ge

    2016-11-16

    Long noncoding RNAs (lncRNAs) have been shown to play essential roles in almost every important biological process through multiple mechanisms. Although the repertoire of human lncRNAs has rapidly expanded, their biological function and regulation remain largely elusive, calling for a systematic and integrative annotation tool. Here we present AnnoLnc ( http://annolnc.cbi.pku.edu.cn ), a one-stop portal for systematically annotating novel human lncRNAs. Based on more than 700 data sources and various tool chains, AnnoLnc enables a systematic annotation covering genomic location, secondary structure, expression patterns, transcriptional regulation, miRNA interaction, protein interaction, genetic association and evolution. An intuitive web interface is available for interactive analysis through both desktops and mobile devices, and programmers can further integrate AnnoLnc into their pipeline through standard JSON-based Web Service APIs. To the best of our knowledge, AnnoLnc is the only web server to provide on-the-fly and systematic annotation for newly identified human lncRNAs. Compared with similar tools, the annotation generated by AnnoLnc covers a much wider spectrum with intuitive visualization. Case studies demonstrate the power of AnnoLnc in not only rediscovering known functions of human lncRNAs but also inspiring novel hypotheses.

  2. New Web Services for Broader Access to National Deep Submergence Facility Data Resources Through the Interdisciplinary Earth Data Alliance

    NASA Astrophysics Data System (ADS)

    Ferrini, V. L.; Grange, B.; Morton, J. J.; Soule, S. A.; Carbotte, S. M.; Lehnert, K.

    2016-12-01

    The National Deep Submergence Facility (NDSF) operates the Human Occupied Vehicle (HOV) Alvin, the Remotely Operated Vehicle (ROV) Jason, and the Autonomous Underwater Vehicle (AUV) Sentry. These vehicles are deployed throughout the global oceans to acquire sensor data and physical samples for a variety of interdisciplinary science programs. As part of the EarthCube Integrative Activity Alliance Testbed Project (ATP), new web services were developed to improve access to existing online NDSF data and metadata resources. These services make use of tools and infrastructure developed by the Interdisciplinary Earth Data Alliance (IEDA) and enable programmatic access to metadata and data resources as well as the development of new service-driven user interfaces. The Alvin Frame Grabber and Jason Virtual Van enable the exploration of frame-grabbed images derived from video cameras on NDSF dives. Metadata available for each image includes time and vehicle position, data from environmental sensors, and scientist-generated annotations, and data are organized and accessible by cruise and/or dive. A new FrameGrabber web service and service-driven user interface were deployed to offer integrated access to these data resources through a single API and allows users to search across content curated in both systems. In addition, a new NDSF Dive Metadata web service and service-driven user interface was deployed to provide consolidated access to basic information about each NDSF dive (e.g. vehicle name, dive ID, location, etc), which is important for linking distributed data resources curated in different data systems.

  3. The Geodetic Seamless Archive Centers Service Layer: A System Architecture for Federating Geodesy Data Repositories

    NASA Astrophysics Data System (ADS)

    McWhirter, J.; Boler, F. M.; Bock, Y.; Jamason, P.; Squibb, M. B.; Noll, C. E.; Blewitt, G.; Kreemer, C. W.

    2010-12-01

    Three geodesy Archive Centers, Scripps Orbit and Permanent Array Center (SOPAC), NASA's Crustal Dynamics Data Information System (CDDIS) and UNAVCO are engaged in a joint effort to define and develop a common Web Service Application Programming Interface (API) for accessing geodetic data holdings. This effort is funded by the NASA ROSES ACCESS Program to modernize the original GPS Seamless Archive Centers (GSAC) technology which was developed in the 1990s. A new web service interface, the GSAC-WS, is being developed to provide uniform and expanded mechanisms through which users can access our data repositories. In total, our respective archives hold tens of millions of files and contain a rich collection of site/station metadata. Though we serve similar user communities, we currently provide a range of different access methods, query services and metadata formats. This leads to a lack of consistency in the userís experience and a duplication of engineering efforts. The GSAC-WS API and its reference implementation in an underlying Java-based GSAC Service Layer (GSL) supports metadata and data queries into site/station oriented data archives. The general nature of this API makes it applicable to a broad range of data systems. The overall goals of this project include providing consistent and rich query interfaces for end users and client programs, the development of enabling technology to facilitate third party repositories in developing these web service capabilities and to enable the ability to perform data queries across a collection of federated GSAC-WS enabled repositories. A fundamental challenge faced in this project is to provide a common suite of query services across a heterogeneous collection of data yet enabling each repository to expose their specific metadata holdings. To address this challenge we are developing a "capabilities" based service where a repository can describe its specific query and metadata capabilities. Furthermore, the architecture of the GSL is based on a model-view paradigm that decouples the underlying data model semantics from particular representations of the data model. This will allow for the GSAC-WS enabled repositories to evolve their service offerings to incorporate new metadata definition formats (e.g., ISO-19115, FGDC, JSON, etc.) and new techniques for accessing their holdings. Building on the core GSAC-WS implementations the project is also developing a federated/distributed query service. This service will seamlessly integrate with the GSAC Service Layer and will support data and metadata queries across a collection of federated GSAC repositories.

  4. Visualizing Cloud Properties and Satellite Imagery: A Tool for Visualization and Information Integration

    NASA Astrophysics Data System (ADS)

    Chee, T.; Nguyen, L.; Smith, W. L., Jr.; Spangenberg, D.; Palikonda, R.; Bedka, K. M.; Minnis, P.; Thieman, M. M.; Nordeen, M.

    2017-12-01

    Providing public access to research products including cloud macro and microphysical properties and satellite imagery are a key concern for the NASA Langley Research Center Cloud and Radiation Group. This work describes a web based visualization tool and API that allows end users to easily create customized cloud product and satellite imagery, ground site data and satellite ground track information that is generated dynamically. The tool has two uses, one to visualize the dynamically created imagery and the other to provide access to the dynamically generated imagery directly at a later time. Internally, we leverage our practical experience with large, scalable application practices to develop a system that has the largest potential for scalability as well as the ability to be deployed on the cloud to accommodate scalability issues. We build upon NASA Langley Cloud and Radiation Group's experience with making real-time and historical satellite cloud product information, satellite imagery, ground site data and satellite track information accessible and easily searchable. This tool is the culmination of our prior experience with dynamic imagery generation and provides a way to build a "mash-up" of dynamically generated imagery and related kinds of information that are visualized together to add value to disparate but related information. In support of NASA strategic goals, our group aims to make as much scientific knowledge, observations and products available to the citizen science, research and interested communities as well as for automated systems to acquire the same information for data mining or other analytic purposes. This tool and the underlying API's provide a valuable research tool to a wide audience both as a standalone research tool and also as an easily accessed data source that can easily be mined or used with existing tools.

  5. Supporting Collaborative Model and Data Service Development and Deployment with DevOps

    NASA Astrophysics Data System (ADS)

    David, O.

    2016-12-01

    Adopting DevOps practices for model service development and deployment enables a community to engage in service-oriented modeling and data management. The Cloud Services Integration Platform (CSIP) developed the last 5 years at Colorado State University provides for collaborative integration of environmental models into scalable model and data services as a micro-services platform with API and deployment infrastructure. Originally developed to support USDA natural resource applications, it proved suitable for a wider range of applications in the environmental modeling domain. While extending its scope and visibility it became apparent community integration and adequate work flow support through the full model development and application cycle drove successful outcomes.DevOps provide best practices, tools, and organizational structures to optimize the transition from model service development to deployment by minimizing the (i) operational burden and (ii) turnaround time for modelers. We have developed and implemented a methodology to fully automate a suite of applications for application lifecycle management, version control, continuous integration, container management, and container scaling to enable model and data service developers in various institutions to collaboratively build, run, deploy, test, and scale services within minutes.To date more than 160 model and data services are available for applications in hydrology (PRMS, Hydrotools, CFA, ESP), water and wind erosion prediction (WEPP, WEPS, RUSLE2), soil quality trends (SCI, STIR), water quality analysis (SWAT-CP, WQM, CFA, AgES-W), stream degradation assessment (SWAT-DEG), hydraulics (cross-section), and grazing management (GRAS). In addition, supporting data services include soil (SSURGO), ecological site (ESIS), climate (CLIGEN, WINDGEN), land management and crop rotations (LMOD), and pesticides (WQM), developed using this workflow automation and decentralized governance.

  6. Near infrared and Raman spectroscopy as Process Analytical Technology tools for the manufacturing of silicone-based drug reservoirs.

    PubMed

    Mantanus, J; Rozet, E; Van Butsele, K; De Bleye, C; Ceccato, A; Evrard, B; Hubert, Ph; Ziémons, E

    2011-08-05

    Using near infrared (NIR) and Raman spectroscopy as PAT tools, 3 critical quality attributes of a silicone-based drug reservoir were studied. First, the Active Pharmaceutical Ingredient (API) homogeneity in the reservoir was evaluated using Raman spectroscopy (mapping): the API distribution within the industrial drug reservoirs was found to be homogeneous while API aggregates were detected in laboratory scale samples manufactured with a non optimal mixing process. Second, the crosslinking process of the reservoirs was monitored at different temperatures with NIR spectroscopy. Conformity tests and Principal Component Analysis (PCA) were performed on the collected data to find out the relation between the temperature and the time necessary to reach the crosslinking endpoints. An agreement was found between the conformity test results and the PCA results. Compared to the conformity test method, PCA had the advantage to discriminate the heating effect from the crosslinking effect occurring together during the monitored process. Therefore the 2 approaches were found to be complementary. Third, based on the HPLC reference method, a NIR model able to quantify the API in the drug reservoir was developed and thoroughly validated. Partial Least Squares (PLS) regression on the calibration set was performed to build prediction models of which the ability to quantify accurately was tested with the external validation set. The 1.2% Root Mean Squared Error of Prediction (RMSEP) of the NIR model indicated the global accuracy of the model. The accuracy profile based on tolerance intervals was used to generate a complete validation report. The 95% tolerance interval calculated on the validation results indicated that each future result will have a relative error below ±5% with a probability of at least 95%. In conclusion, 3 critical quality attributes of silicone-based drug reservoirs were quickly and efficiently evaluated by NIR and Raman spectroscopy. Copyright © 2011 Elsevier B.V. All rights reserved.

  7. Development and in-line validation of a Process Analytical Technology to facilitate the scale up of coating processes.

    PubMed

    Wirges, M; Funke, A; Serno, P; Knop, K; Kleinebudde, P

    2013-05-05

    Incorporation of an active pharmaceutical ingredient (API) into the coating layer of film-coated tablets is a method mainly used to formulate fixed-dose combinations. Uniform and precise spray-coating of an API represents a substantial challenge, which could be overcome by applying Raman spectroscopy as process analytical tool. In pharmaceutical industry, Raman spectroscopy is still mainly used as a bench top laboratory analytical method and usually not implemented in the production process. Concerning the application in the production process, a lot of scientific approaches stop at the level of feasibility studies and do not manage the step to production scale and process applications. The present work puts the scale up of an active coating process into focus, which is a step of highest importance during the pharmaceutical development. Active coating experiments were performed at lab and production scale. Using partial least squares (PLS), a multivariate model was constructed by correlating in-line measured Raman spectral data with the coated amount of API. By transferring this model, being implemented for a lab scale process, to a production scale process, the robustness of this analytical method and thus its applicability as a Process Analytical Technology (PAT) tool for the correct endpoint determination in pharmaceutical manufacturing could be shown. Finally, this method was validated according to the European Medicine Agency (EMA) guideline with respect to the special requirements of the applied in-line model development strategy. Copyright © 2013 Elsevier B.V. All rights reserved.

  8. A Process Analytical Technology (PAT) approach to control a new API manufacturing process: development, validation and implementation.

    PubMed

    Schaefer, Cédric; Clicq, David; Lecomte, Clémence; Merschaert, Alain; Norrant, Edith; Fotiadu, Frédéric

    2014-03-01

    Pharmaceutical companies are progressively adopting and introducing Process Analytical Technology (PAT) and Quality-by-Design (QbD) concepts promoted by the regulatory agencies, aiming the building of the quality directly into the product by combining thorough scientific understanding and quality risk management. An analytical method based on near infrared (NIR) spectroscopy was developed as a PAT tool to control on-line an API (active pharmaceutical ingredient) manufacturing crystallization step during which the API and residual solvent contents need to be precisely determined to reach the predefined seeding point. An original methodology based on the QbD principles was designed to conduct the development and validation of the NIR method and to ensure that it is fitted for its intended use. On this basis, Partial least squares (PLS) models were developed and optimized using chemometrics methods. The method was fully validated according to the ICH Q2(R1) guideline and using the accuracy profile approach. The dosing ranges were evaluated to 9.0-12.0% w/w for the API and 0.18-1.50% w/w for the residual methanol. As by nature the variability of the sampling method and the reference method are included in the variability obtained for the NIR method during the validation phase, a real-time process monitoring exercise was performed to prove its fit for purpose. The implementation of this in-process control (IPC) method on the industrial plant from the launch of the new API synthesis process will enable automatic control of the final crystallization step in order to ensure a predefined quality level of the API. In addition, several valuable benefits are expected including reduction of the process time, suppression of a rather difficult sampling and tedious off-line analyses. © 2013 Published by Elsevier B.V.

  9. Flight behaviour of honey bee (Apis mellifera) workers is altered by initial infections of the fungal parasite Nosema apis

    PubMed Central

    Dosselli, Ryan; Grassl, Julia; Carson, Andrew; Simmons, Leigh W.; Baer, Boris

    2016-01-01

    Honey bees (Apis mellifera) host a wide range of parasites, some being known contributors towards dramatic colony losses as reported over recent years. To counter parasitic threats, honey bees possess effective immune systems. Because immune responses are predicted to cause substantial physiological costs for infected individuals, they are expected to trade off with other life history traits that ultimately affect the performance and fitness of the entire colony. Here, we tested whether the initial onset of an infection negatively impacts the flight behaviour of honey bee workers, which is an energetically demanding behaviour and a key component of foraging activities. To do this, we infected workers with the widespread fungal pathogen Nosema apis, which is recognised and killed by the honey bee immune system. We compared their survival and flight behaviour with non-infected individuals from the same cohort and colony using radio frequency identification tags (RFID). We found that over a time frame of four days post infection, Nosema did not increase mortality but workers quickly altered their flight behaviour and performed more flights of shorter duration. We conclude that parasitic infections influence foraging activities, which could reduce foraging ranges of colonies and impact their ability to provide pollination services. PMID:27827404

  10. Flight behaviour of honey bee (Apis mellifera) workers is altered by initial infections of the fungal parasite Nosema apis.

    PubMed

    Dosselli, Ryan; Grassl, Julia; Carson, Andrew; Simmons, Leigh W; Baer, Boris

    2016-11-09

    Honey bees (Apis mellifera) host a wide range of parasites, some being known contributors towards dramatic colony losses as reported over recent years. To counter parasitic threats, honey bees possess effective immune systems. Because immune responses are predicted to cause substantial physiological costs for infected individuals, they are expected to trade off with other life history traits that ultimately affect the performance and fitness of the entire colony. Here, we tested whether the initial onset of an infection negatively impacts the flight behaviour of honey bee workers, which is an energetically demanding behaviour and a key component of foraging activities. To do this, we infected workers with the widespread fungal pathogen Nosema apis, which is recognised and killed by the honey bee immune system. We compared their survival and flight behaviour with non-infected individuals from the same cohort and colony using radio frequency identification tags (RFID). We found that over a time frame of four days post infection, Nosema did not increase mortality but workers quickly altered their flight behaviour and performed more flights of shorter duration. We conclude that parasitic infections influence foraging activities, which could reduce foraging ranges of colonies and impact their ability to provide pollination services.

  11. Dataflow Integration and Simulation Techniques for DSP System Design Tools

    DTIC Science & Technology

    2007-01-01

    Lebak, M. Richards , and D. Campbell, “VSIPL: An object-based open standard API for vector, signal, and image processing,” in Proceedings of the...Inc., document Version 0.98a. [56] P. Marwedel and G. Goossens , Eds., Code Generation for Embedded Processors. Kluwer Academic Publishers, 1995. [57

  12. Growth Index: A Powerful Tool for School Improvement

    ERIC Educational Resources Information Center

    Wiseman, Perry; Thomas, Kimberly

    2011-01-01

    This article outlines a formula called Growth Index (GI), which is designed to measure student growth within all proficiency levels year after year, and track classroom or schoolwide growth. This growth should be considered as one of several dimensions that constitute principal and teacher effectiveness. Unlike Academic Performance Index (API) and…

  13. MyGeoHub: A Collaborative Geospatial Research and Education Platform

    NASA Astrophysics Data System (ADS)

    Kalyanam, R.; Zhao, L.; Biehl, L. L.; Song, C. X.; Merwade, V.; Villoria, N.

    2017-12-01

    Scientific research is increasingly collaborative and globally distributed; research groups now rely on web-based scientific tools and data management systems to simplify their day-to-day collaborative workflows. However, such tools often lack seamless interfaces, requiring researchers to contend with manual data transfers, annotation and sharing. MyGeoHub is a web platform that supports out-of-the-box, seamless workflows involving data ingestion, metadata extraction, analysis, sharing and publication. MyGeoHub is built on the HUBzero cyberinfrastructure platform and adds general-purpose software building blocks (GABBs), for geospatial data management, visualization and analysis. A data management building block iData, processes geospatial files, extracting metadata for keyword and map-based search while enabling quick previews. iData is pervasive, allowing access through a web interface, scientific tools on MyGeoHub or even mobile field devices via a data service API. GABBs includes a Python map library as well as map widgets that in a few lines of code, generate complete geospatial visualization web interfaces for scientific tools. GABBs also includes powerful tools that can be used with no programming effort. The GeoBuilder tool provides an intuitive wizard for importing multi-variable, geo-located time series data (typical of sensor readings, GPS trackers) to build visualizations supporting data filtering and plotting. MyGeoHub has been used in tutorials at scientific conferences and educational activities for K-12 students. MyGeoHub is also constantly evolving; the recent addition of Jupyter and R Shiny notebook environments enable reproducible, richly interactive geospatial analyses and applications ranging from simple pre-processing to published tools. MyGeoHub is not a monolithic geospatial science gateway, instead it supports diverse needs ranging from just a feature-rich data management system, to complex scientific tools and workflows.

  14. The Virtual Space Physics Observatory: Quick Access to Data and Tools

    NASA Technical Reports Server (NTRS)

    Cornwell, Carl; Roberts, D. Aaron; McGuire, Robert E.

    2006-01-01

    The Virtual Space Physics Observatory (VSPO; see http://vspo.gsfc.nasa.gov) has grown to provide a way to find and access about 375 data products and services from over 100 spacecraft/observatories in space and solar physics. The datasets are mainly chosen to be the most requested, and include most of the publicly available data products from operating NASA Heliophysics spacecraft as well as from solar observatories measuring across the frequency spectrum. Service links include a "quick orbits" page that uses SSCWeb Web Services to provide a rapid answer to questions such as "What spacecraft were in orbit in July 1992?" and "Where were Geotail, Cluster, and Polar on 2 June 2001?" These queries are linked back to the data search page. The VSPO interface provides many ways of looking for data based on terms used in a registry of resources using the SPASE Data Model that will be the standard for Heliophysics Virtual Observatories. VSPO itself is accessible via an API that allows other applications to use it as a Web Service; this has been implemented in one instance using the ViSBARD visualization program. The VSPO will become part of the Space Physics Data Facility, and will continue to expand its access to data. A challenge for all VOs will be to provide uniform access to data at the variable level, and we will be addressing this question in a number of ways.

  15. Integrating Containers in the CERN Private Cloud

    NASA Astrophysics Data System (ADS)

    Noel, Bertrand; Michelino, Davide; Velten, Mathieu; Rocha, Ricardo; Trigazis, Spyridon

    2017-10-01

    Containers remain a hot topic in computing, with new use cases and tools appearing every day. Basic functionality such as spawning containers seems to have settled, but topics like volume support or networking are still evolving. Solutions like Docker Swarm, Kubernetes or Mesos provide similar functionality but target different use cases, exposing distinct interfaces and APIs. The CERN private cloud is made of thousands of nodes and users, with many different use cases. A single solution for container deployment would not cover every one of them, and supporting multiple solutions involves repeating the same process multiple times for integration with authentication services, storage services or networking. In this paper we describe OpenStack Magnum as the solution to offer container management in the CERN cloud. We will cover its main functionality and some advanced use cases using Docker Swarm and Kubernetes, highlighting some relevant differences between the two. We will describe the most common use cases in HEP and how we integrated popular services like CVMFS or AFS in the most transparent way possible, along with some limitations found. Finally we will look into ongoing work on advanced scheduling for both Swarm and Kubernetes, support for running batch like workloads and integration of container networking technologies with the CERN infrastructure.

  16. Preliminary characterization of digestive enzymes in freshwater mussels

    USGS Publications Warehouse

    Sauey, Blake W.; Amberg, Jon J.; Cooper, Scott T.; Grunwald, Sandra K.; Newton, Teresa J.; Haro, Roger J.

    2015-01-01

    Resource managers lack an effective chemical tool to control the invasive zebra mussel Dreissena polymorpha. Zebra mussels clog water intakes for hydroelectric companies, harm unionid mussel species, and are believed to be a reservoir of avian botulism. Little is known about the digestive physiology of zebra mussels and unionid mussels. The enzymatic profile of the digestive glands of zebra mussels and native threeridge (Amblema plicata) and plain pocketbook mussels (Lampsilis cardium) are characterized using a commercial enzyme kit, api ZYM, and validated the kit with reagent-grade enzymes. A linear correlation was shown for only one of nineteen enzymes, tested between the api ZYM kit and a specific enzyme kit. Thus, the api ZYM kit should only be used to make general comparisons of enzyme presence and to observe trends in enzyme activities. Enzymatic trends were seen in the unionid mussel species, but not in zebra mussels sampled 32 days apart from the same location. Enzymatic classes, based on substrate, showed different trends, with proteolytic and phospholytic enzymes having the most change in relative enzyme activity.

  17. Parametric Modelling of As-Built Beam Framed Structure in Bim Environment

    NASA Astrophysics Data System (ADS)

    Yang, X.; Koehl, M.; Grussenmeyer, P.

    2017-02-01

    A complete documentation and conservation of a historic timber roof requires the integration of geometry modelling, attributional and dynamic information management and results of structural analysis. Recently developed as-built Building Information Modelling (BIM) technique has the potential to provide a uniform platform, which provides possibility to integrate the traditional geometry modelling, parametric elements management and structural analysis together. The main objective of the project presented in this paper is to develop a parametric modelling tool for a timber roof structure whose elements are leaning and crossing beam frame. Since Autodesk Revit, as the typical BIM software, provides the platform for parametric modelling and information management, an API plugin, able to automatically create the parametric beam elements and link them together with strict relationship, was developed. The plugin under development is introduced in the paper, which can obtain the parametric beam model via Autodesk Revit API from total station points and terrestrial laser scanning data. The results show the potential of automatizing the parametric modelling by interactive API development in BIM environment. It also integrates the separate data processing and different platforms into the uniform Revit software.

  18. On Parallelizing Single Dynamic Simulation Using HPC Techniques and APIs of Commercial Software

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Diao, Ruisheng; Jin, Shuangshuang; Howell, Frederic

    Time-domain simulations are heavily used in today’s planning and operation practices to assess power system transient stability and post-transient voltage/frequency profiles following severe contingencies to comply with industry standards. Because of the increased modeling complexity, it is several times slower than real time for state-of-the-art commercial packages to complete a dynamic simulation for a large-scale model. With the growing stochastic behavior introduced by emerging technologies, power industry has seen a growing need for performing security assessment in real time. This paper presents a parallel implementation framework to speed up a single dynamic simulation by leveraging the existing stability model librarymore » in commercial tools through their application programming interfaces (APIs). Several high performance computing (HPC) techniques are explored such as parallelizing the calculation of generator current injection, identifying fast linear solvers for network solution, and parallelizing data outputs when interacting with APIs in the commercial package, TSAT. The proposed method has been tested on a WECC planning base case with detailed synchronous generator models and exhibits outstanding scalable performance with sufficient accuracy.« less

  19. The RNASeq-er API-a gateway to systematically updated analysis of public RNA-seq data.

    PubMed

    Petryszak, Robert; Fonseca, Nuno A; Füllgrabe, Anja; Huerta, Laura; Keays, Maria; Tang, Y Amy; Brazma, Alvis

    2017-07-15

    The exponential growth of publicly available RNA-sequencing (RNA-Seq) data poses an increasing challenge to researchers wishing to discover, analyse and store such data, particularly those based in institutions with limited computational resources. EMBL-EBI is in an ideal position to address these challenges and to allow the scientific community easy access to not just raw, but also processed RNA-Seq data. We present a Web service to access the results of a systematically and continually updated standardized alignment as well as gene and exon expression quantification of all public bulk (and in the near future also single-cell) RNA-Seq runs in 264 species in European Nucleotide Archive, using Representational State Transfer. The RNASeq-er API (Application Programming Interface) enables ontology-powered search for and retrieval of CRAM, bigwig and bedGraph files, gene and exon expression quantification matrices (Fragments Per Kilobase Of Exon Per Million Fragments Mapped, Transcripts Per Million, raw counts) as well as sample attributes annotated with ontology terms. To date over 270 00 RNA-Seq runs in nearly 10 000 studies (1PB of raw FASTQ data) in 264 species in ENA have been processed and made available via the API. The RNASeq-er API can be accessed at http://www.ebi.ac.uk/fg/rnaseq/api . The commands used to analyse the data are available in supplementary materials and at https://github.com/nunofonseca/irap/wiki/iRAP-single-library . rnaseq@ebi.ac.uk ; rpetry@ebi.ac.uk. Supplementary data are available at Bioinformatics online. © The Author(s) 2017. Published by Oxford University Press.

  20. Services, Perspective and Directions of the Space Physics Data Facility

    NASA Technical Reports Server (NTRS)

    McGuire, Robert E.; Bilitza, Dieter; Candey, Reine A.; Chimiak, Reine A.; Cooper, John F.; Fung, Shing F.; Harris, Bernard T.; Johnson, Rita C.; King, Joseph H.; Kovalick, Tamara; hide

    2008-01-01

    The multi-mission data and orbit services of NASA's Space Physics Data Facility (SPDF) project offer unique capabilities supporting science of the Heliophysics Great Observatory and that are highly complementary to other services now evolving in the international heliophysics data environment. The VSPO (Virtual Space Physics Observatory) service is an active portal to a wide rage of distributed data sources. CDAWeb (Coordinated Data Analysis Web) offers plots, listings and file downloads for current data from many missions across the boundaries of missions and instrument types. CDAWeb now includes extensive new data from STEREO and THEMIS, plus new ROCSAT IPEI data, the latest data from all four TIMED instruments and high-resolution data from all DE-2 experiments. SSCWeb, Helioweb and out 3D Animated Orbit Viewer (TIPSOD) provide position data and identification of spacecraft and ground conjunctions. OMNI Web, with its new extension to 1- and 5-minute resolution, provides interplanetary parameters at the Earth's bow shock. SPDF maintains NASA's CDF (Common Data Format) standard and a range of associated tools including format translation services. These capabilities are all now available through web services based APIs, one element in SPDF's ongoing work to enable heliophysics community development of Virtual discipline Observatories (e.g. VITMO). We will demonstrate out latest data and capabilities, review the lessons we continue to learn in what science users need and value in this class of services, and discuss out current thinking to the future role and appropriate focus of the SPDF effort in the evolving and increasingly distributed heliophysics data environment.

  1. A Million Cancer Genome Warehouse

    DTIC Science & Technology

    2012-11-20

    Software, Strawberry Canyon, 2012. 25 Units (GPUs) without any changes needed to the client applications. ● Service-level APIs are designed to... Strawberry Canyon, 2012. 62 Patterson, D. For better or worse, benchmarks shape a field: technical perspective, Communications of the ACM, v.55 n.7

  2. jTraML: an open source Java API for TraML, the PSI standard for sharing SRM transitions.

    PubMed

    Helsens, Kenny; Brusniak, Mi-Youn; Deutsch, Eric; Moritz, Robert L; Martens, Lennart

    2011-11-04

    We here present jTraML, a Java API for the Proteomics Standards Initiative TraML data standard. The library provides fully functional classes for all elements specified in the TraML XSD document, as well as convenient methods to construct controlled vocabulary-based instances required to define SRM transitions. The use of jTraML is demonstrated via a two-way conversion tool between TraML documents and vendor specific files, facilitating the adoption process of this new community standard. The library is released as open source under the permissive Apache2 license and can be downloaded from http://jtraml.googlecode.com . TraML files can also be converted online at http://iomics.ugent.be/jtraml .

  3. Analyzing CRISM hyperspectral imagery using PlanetServer.

    NASA Astrophysics Data System (ADS)

    Figuera, Ramiro Marco; Pham Huu, Bang; Minin, Mikhail; Flahaut, Jessica; Halder, Anik; Rossi, Angelo Pio

    2017-04-01

    Mineral characterization of planetary surfaces bears great importance for space exploration. In order to perform it, orbital hyperspectral imagery is widely used. In our research we use Compact Reconnaissance Imaging Spectrometer for Mars (CRISM) [1] TRDR L observations with a spectral range of 1 to 4 µm. PlanetServer comprises a server, a web client and a Python client/API. The server side uses the Array DataBase Management System (DBMS) Raster Data Manager (Rasdaman) Community Edition [2]. OGC standards such as the Web Coverage Processing Service (WCPS) [3], an SQL-like language capable to query information along the image cube, are implemented in the PetaScope component [4]. The client side uses NASA's Web World Wind [5] allowing the user to access the data in an intuitive way. The client consists of a globe where all cubes are deployed, a main menu where projections, base maps and RGB combinations are provided, and a plot dock where the spectral information is shown. The RGB combinator tool allows to do band combination such as the CRISM products [6] using WCPS. The spectral information is retrieved using WCPS and shown in the plot dock/widget. The USGS splib06a library [7] is available to compare CRISM vs. laboratory spectra. The Python API provides an environment to create RGB combinations that can be embedded into existing pipelines. All employed libraries and tools are open source and can be easily adapted to other datasets. PlanetServer stands as a promising tool for spectral analysis on planetary bodies. M3/Moon and OMEGA datasets will be soon available. [1] S. Murchie et al., "Compact Connaissance Imaging Spectrometer for Mars (CRISM) on Mars Reconnaissance Orbiter (MRO)," J. Geophys. Res. E Planets,2007. [2] P. Baumann, A. Dehmel, P. Furtado, R. Ritsch, and N. Widmann, "The multidimensional database system RasDaMan," ACM SIGMOD Rec., vol. 27, no. 2, pp. 575-577, Jun. 1998. [3] P. Baumann, "The OGC web coverage processing service (WCPS) standard," Geoinformatica, vol. 14, no. 4, Jul. 2010. [4] A. Aiordǎchioaie and P. Baumann, "PetaScope: An open-source implementation of the OGC WCS Geo service standards suite," Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 6187 LNCS, pp. 160-168, Jun. 2010. [5] P. Hogan, C. Maxwell, R. Kim, and T. Gaskins, "World Wind 3D Earth Viewing," Apr. 2007. [6] C. E. Viviano-Beck et al., "Revised CRISM spectral parameters and summary products based on the currently detected mineral diversity on Mars," J. Geophys. Res. E Planets, vol. 119, no. 6, pp. 1403-1431, Jun. 2014. [7] R. N. Clark et al., "USGS digital spectral library splib06a: U.S. Geological Survey, Digital Data Series 231," 2007. [Online]. Available: http://speclab.cr.usgs.gov/spectral.lib06.

  4. Expanding the use of Scientific Data through Maps and Apps

    NASA Astrophysics Data System (ADS)

    Shrestha, S. R.; Zimble, D. A.; Herring, D.; Halpert, M.

    2014-12-01

    The importance of making scientific data more available can't be overstated. There is a wealth of useful scientific data available and demand for this data is only increasing; however, applying scientific data towards practical uses poses several technical challenges. These challenges can arise from difficulty in handling the data due largely to 1) the complexity, variety and volume of scientific data and 2) applying and operating the techniques and tools needed to visualize and analyze the data. As a result, the combined knowledge required to take advantage of these data requires highly specialized skill sets that in total, limit the ability of scientific data from being used in more practical day-to-day decision making activities. While these challenges are daunting, information technologies do exist that can help mitigate some of these issues. Many organizations for years have already been enjoying the benefits of modern service oriented architectures (SOAs) for everyday enterprise tasks. We can use this approach to modernize how we share and access our scientific data where much of the specialized tools and techniques needed to handle and present scientific data can be automated and executed by servers and done so in an appropriate way. We will discuss and show an approach for preparing file based scientific data (e.g. GRIB, netCDF) for use in standard based scientific web services. These scientific web services are able to encapsulate the logic needed to handle and describe scientific data through a variety of service types including, image, map, feature, geoprocessing, and their respective service methods. By combining these types of services and leveraging well-documented and modern web development APIs, we can afford to focus our attention on the design and development of user-friendly maps and apps. Our scenario will include developing online maps through these services by integrating various forecast data from the Climate Forecast System (CFSv2). This presentation showcases a collaboration between the National Oceanic and Atmospheric Administration's (NOAA) Climate.gov portal, Climate Prediction Center and Esri, Inc. on the implementation of the ArcGIS platform, which is aimed at helping modernize scientific data access through a service oriented architecture.

  5. The Philosophy of User Interfaces in HELIO and the Importance of CASSIS

    NASA Astrophysics Data System (ADS)

    Bonnin, X.; Aboudarham, J.; Renié, C.; Csillaghy, A.; Messerotti, M.; Bentley, R. D.

    2012-09-01

    HELIO is a European project funded under FP7 (Project No. 238969). One of its goals as a Heliospheric Virtual Observatory is to provide an easy access to many datasets scattered all over the world, in the fields of Solar physics, Heliophysics, and Planetary magnetospheres. The efficiency of such a tool is very much related to the quality of the user interface. HELIO infrastructure is based on a Service Oriented Architecture (SOA), regrouping a network of standalone components, which allows four main types of interfaces: - HELIO Front End (HFE) is a browser-based user interface, which offers a centralized access to the HELIO main functionalities. Especially, it provides the possibility to reach data directly, or to refine selection by determination of observing characteristics, such as which instrument was observing at that time, which instrument was at this location, etc. - Many services/components provide their own standalone graphical user interface. While one can directly access individually each of these interfaces, they can also be connected together. - Most services also provide direct access for any tools through a public interface. A small java library, called Java API, simplifies this access by providing client stubs for services and shields the user from security, discovery and failover issues. - Workflows capabilities are available in HELIO, allowing complex combination of queries over several services. We want the user to be able to navigate easily, at his needs, through the various interfaces, and possibly use a specific one in order to make much-dedicated queries. We will also emphasize the importance of the CASSIS project (Coordination Action for the integration of Solar System Infrastructure and Science) in encouraging the interoperability necessary to undertake scientific studies that span disciplinary boundaries. If related projects follow the guidelines being developed by CASSIS then using external resources with HELIO will be greatly simplified.

  6. Asynchronous Object Storage with QoS for Scientific and Commercial Big Data

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Brim, Michael J; Dillow, David A; Oral, H Sarp

    2013-01-01

    This paper presents our design for an asynchronous object storage system intended for use in scientific and commercial big data workloads. Use cases from the target workload do- mains are used to motivate the key abstractions used in the application programming interface (API). The architecture of the Scalable Object Store (SOS), a prototype object stor- age system that supports the API s facilities, is presented. The SOS serves as a vehicle for future research into scalable and resilient big data object storage. We briefly review our research into providing efficient storage servers capable of providing quality of service (QoS) contractsmore » relevant for big data use cases.« less

  7. Cloud-Enabled Climate Analytics-as-a-Service using Reanalysis data: A case study.

    NASA Astrophysics Data System (ADS)

    Nadeau, D.; Duffy, D.; Schnase, J. L.; McInerney, M.; Tamkin, G.; Potter, G. L.; Thompson, J. H.

    2014-12-01

    The NASA Center for Climate Simulation (NCCS) maintains advanced data capabilities and facilities that allow researchers to access the enormous volume of data generated by weather and climate models. The NASA Climate Model Data Service (CDS) and the NCCS are merging their efforts to provide Climate Analytics-as-a-Service for the comparative study of the major reanalysis projects: ECMWF ERA-Interim, NASA/GMAO MERRA, NOAA/NCEP CFSR, NOAA/ESRL 20CR, JMA JRA25, and JRA55. These reanalyses have been repackaged to netCDF4 file format following the CMIP5 Climate and Forecast (CF) metadata convention prior to be sequenced into the Hadoop Distributed File System ( HDFS ). A small set of operations that represent a common starting point in many analysis workflows was then created: min, max, sum, count, variance and average. In this example, Reanalysis data exploration was performed with the use of Hadoop MapReduce and accessibility was achieved using the Climate Data Service(CDS) application programming interface (API) created at NCCS. This API provides a uniform treatment of large amount of data. In this case study, we have limited our exploration to 2 variables, temperature and precipitation, using 3 operations, min, max and avg and using 30-year of Reanalysis data for 3 regions of the world: global, polar, subtropical.

  8. Investigation of drug-excipient compatibility using rheological and thermal tools

    NASA Astrophysics Data System (ADS)

    Trivedi, Maitri R.

    HYPOTHESIS: We plan to investigate a different approach to evaluate drug-excipient physical compatibility using rheological and thermal tools as opposed to commonly used chemical techniques in pharmaceutical industry. This approach offers practical solutions to routinely associated problems arising with API's and commonly used hydrates forms of excipients. ABSTRACT: Drug-Excipient compatibility studies are an important aspect of pre-formulation and formulation development in pharmaceutical research and development. Various approaches have been used in pharmaceutical industry including use of thermal analysis and quantitative assessment of drug-excipient mixtures after keeping the samples under stress environment depending upon the type of formulation. In an attempt to provide better understanding of such compatibility aspect of excipients with different properties of API, various rheological and thermal studies were conducted on binary mixtures of excipients which exist in different hydrates. Dibasic Calcium Phosphate (DCP, anhydrous and dihydrate forms) and Lactose (Lac, anhydrous and monohydrate) were selected with cohesive API's (Acetaminophen and Aspirin). Binary mixtures of DCP and Lac were prepared by addition of 0% w/w to 50% w/w of the API into each powder blend. Rheological and thermal aspects were considered using different approaches such as powder rheometer, rotational shear cell and traditional rheometery approaches like angle of repose (AOR), hausner's ratio (HR) and cares index (CI). Thermal analysis was conducted using modulated differential scanning calorimetry (MDSC) and thermal effusivity. The data suggested that the powder rheometer showed distinctive understanding in the flowability behavior of binary mixtures with addition of increasing proportion of API's than traditional approaches. Thermal approaches revealed the potential interaction of water of crystallization DCP-D with the API (APAP) while such interactions were absent in DCP-A, while in case of Lac-M and Lac-A, interaction with water of crystallization were not present. Binary mixtures prepared with DCP-D were better flowable while blends with DCP-A were better in stability (physical), compressibility and permeability. Similarly binary mixtures prepared with Lac-M were better flowable and stable in physical compatibility as compared to Lac-A. Lac-A were better in compressibility and permeability. Second part of these research included understanding the powder behavior from wet granulation point of view. Wet granulation includes the formation of agglomerates with powders to form granules in order to have better flowability, content uniformity and compressibility of granular mass. End point determination of powders involving change in powder energies and compressibility, permeability along with thermal analyses were conducted. The effects of water of crystallization on end point determination was studied and based on which overall effects on drug-excipient compatibility using different hydrate forms of excipients were evaluated.

  9. bioalcidae, samjs and vcffilterjs: object-oriented formatters and filters for bioinformatics files.

    PubMed

    Lindenbaum, Pierre; Redon, Richard

    2018-04-01

    Reformatting and filtering bioinformatics files are common tasks for bioinformaticians. Standard Linux tools and specific programs are usually used to perform such tasks but there is still a gap between using these tools and the programming interface of some existing libraries. In this study, we developed a set of tools namely bioalcidae, samjs and vcffilterjs that reformat or filter files using a JavaScript engine or a pure java expression and taking advantage of the java API for high-throughput sequencing data (htsjdk). https://github.com/lindenb/jvarkit. pierre.lindenbaum@univ-nantes.fr.

  10. High sample throughput genotyping for estimating C-lineage introgression in the dark honeybee: an accurate and cost-effective SNP-based tool.

    PubMed

    Henriques, Dora; Browne, Keith A; Barnett, Mark W; Parejo, Melanie; Kryger, Per; Freeman, Tom C; Muñoz, Irene; Garnery, Lionel; Highet, Fiona; Jonhston, J Spencer; McCormack, Grace P; Pinto, M Alice

    2018-06-04

    The natural distribution of the honeybee (Apis mellifera L.) has been changed by humans in recent decades to such an extent that the formerly widest-spread European subspecies, Apis mellifera mellifera, is threatened by extinction through introgression from highly divergent commercial strains in large tracts of its range. Conservation efforts for A. m. mellifera are underway in multiple European countries requiring reliable and cost-efficient molecular tools to identify purebred colonies. Here, we developed four ancestry-informative SNP assays for high sample throughput genotyping using the iPLEX Mass Array system. Our customized assays were tested on DNA from individual and pooled, haploid and diploid honeybee samples extracted from different tissues using a diverse range of protocols. The assays had a high genotyping success rate and yielded accurate genotypes. Performance assessed against whole-genome data showed that individual assays behaved well, although the most accurate introgression estimates were obtained for the four assays combined (117 SNPs). The best compromise between accuracy and genotyping costs was achieved when combining two assays (62 SNPs). We provide a ready-to-use cost-effective tool for accurate molecular identification and estimation of introgression levels to more effectively monitor and manage A. m. mellifera conservatories.

  11. Final Technical Report - Center for Technology for Advanced Scientific Component Software (TASCS)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sussman, Alan

    2014-10-21

    This is a final technical report for the University of Maryland work in the SciDAC Center for Technology for Advanced Scientific Component Software (TASCS). The Maryland work focused on software tools for coupling parallel software components built using the Common Component Architecture (CCA) APIs. Those tools are based on the Maryland InterComm software framework that has been used in multiple computational science applications to build large-scale simulations of complex physical systems that employ multiple separately developed codes.

  12. The research and implementation of coalfield spontaneous combustion of carbon emission WebGIS based on Silverlight and ArcGIS server

    NASA Astrophysics Data System (ADS)

    Zhu, Z.; Bi, J.; Wang, X.; Zhu, W.

    2014-02-01

    As an important sub-topic of the natural process of carbon emission data public information platform construction, coalfield spontaneous combustion of carbon emission WebGIS system has become an important study object. In connection with data features of coalfield spontaneous combustion carbon emissions (i.e. a wide range of data, which is rich and complex) and the geospatial characteristics, data is divided into attribute data and spatial data. Based on full analysis of the data, completed the detailed design of the Oracle database and stored on the Oracle database. Through Silverlight rich client technology and the expansion of WCF services, achieved the attribute data of web dynamic query, retrieval, statistical, analysis and other functions. For spatial data, we take advantage of ArcGIS Server and Silverlight-based API to invoke GIS server background published map services, GP services, Image services and other services, implemented coalfield spontaneous combustion of remote sensing image data and web map data display, data analysis, thematic map production. The study found that the Silverlight technology, based on rich client and object-oriented framework for WCF service, can efficiently constructed a WebGIS system. And then, combined with ArcGIS Silverlight API to achieve interactive query attribute data and spatial data of coalfield spontaneous emmission, can greatly improve the performance of WebGIS system. At the same time, it provided a strong guarantee for the construction of public information on China's carbon emission data.

  13. Microservices for systematic profiling and monitoring of the refactoring process at the LHCb experiment

    NASA Astrophysics Data System (ADS)

    Mazurov, Alexander; Couturier, Ben; Popov, Dmitry; Farley, Nathanael

    2017-10-01

    Any time you modify an implementation within a program, change compiler version or operating system, you should also do regression testing. You can do regression testing by rerunning existing tests against the changes to determine whether this breaks anything that worked prior to the change and by writing new tests where necessary. At LHCb we have a huge codebase which is maintained by many people and can be run within different setups. Such situations lead to the crucial necessity to guide refactoring with a central profiling system that helps to run tests and find the impact of changes. In our work we present a software architecture and tools for running a profiling system. This system is responsible for systematically running regression tests, collecting and comparing results of these tests so changes between different setups can be observed and reported. The main feature of our solution is that it is based on a microservices architecture. Microservices break a large project into loosely coupled modules, which communicate with each other through simple APIs. Such modular architectural style helps us to avoid general pitfalls of monolithic architectures such as hard to understand a codebase as well as maintaining a large codebase and ineffective scalability. Our solution also allows to escape a complexity of microservices deployment process by using software containers and services management tools. Containers and service managers let us quickly deploy linked modules in development, production or in any other environments. Most of the developed modules are generic which means that the proposed architecture and tools can be used not only in LHCb but adopted for other experiments and companies.

  14. An Approach to Function Annotation for Proteins of Unknown Function (PUFs) in the Transcriptome of Indian Mulberry.

    PubMed

    Dhanyalakshmi, K H; Naika, Mahantesha B N; Sajeevan, R S; Mathew, Oommen K; Shafi, K Mohamed; Sowdhamini, Ramanathan; N Nataraja, Karaba

    2016-01-01

    The modern sequencing technologies are generating large volumes of information at the transcriptome and genome level. Translation of this information into a biological meaning is far behind the race due to which a significant portion of proteins discovered remain as proteins of unknown function (PUFs). Attempts to uncover the functional significance of PUFs are limited due to lack of easy and high throughput functional annotation tools. Here, we report an approach to assign putative functions to PUFs, identified in the transcriptome of mulberry, a perennial tree commonly cultivated as host of silkworm. We utilized the mulberry PUFs generated from leaf tissues exposed to drought stress at whole plant level. A sequence and structure based computational analysis predicted the probable function of the PUFs. For rapid and easy annotation of PUFs, we developed an automated pipeline by integrating diverse bioinformatics tools, designated as PUFs Annotation Server (PUFAS), which also provides a web service API (Application Programming Interface) for a large-scale analysis up to a genome. The expression analysis of three selected PUFs annotated by the pipeline revealed abiotic stress responsiveness of the genes, and hence their potential role in stress acclimation pathways. The automated pipeline developed here could be extended to assign functions to PUFs from any organism in general. PUFAS web server is available at http://caps.ncbs.res.in/pufas/ and the web service is accessible at http://capservices.ncbs.res.in/help/pufas.

  15. How to Sustainably Increase Students' Willingness to Protect Pollinators

    ERIC Educational Resources Information Center

    Schönfelder, Mona L.; Bogner, Franz X.

    2018-01-01

    The current loss of biodiversity requires efforts to increase awareness of pollinator conservation. An important tool is education which often uses the honeybee ("Apis mellifera") as an exemplary organism to reach this goal. Any successful module needs to focus on reducing the perceived danger associated with fear, in order to support…

  16. Adapting Stanford's Chronic Disease Self-Management Program to Hawaii's Multicultural Population

    ERIC Educational Resources Information Center

    Tomioka, Michiyo; Braun, Kathryn L.; Compton, Merlita; Tanoue, Leslie

    2012-01-01

    Purpose of the Study: Stanford's Chronic Disease Self-Management Program (CDSMP) has been proven to increase patients' ability to manage distress. We describe how we replicated CDSMP in Asian and Pacific Islander (API) communities. Design and Methods: We used the "track changes" tool to deconstruct CDSMP into its various components…

  17. SimEngine v. 1.0

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Le, Hai D.

    2017-03-02

    SimEngine provides the core functionalities and components that are key to the development of discrete event simulation tools. These include events, activities, event queues, random number generators, and basic result tracking classes. SimEngine was designed for high performance, integrates seamlessly into any Microsoft .Net development environment, and provides a flexible API for simulation developers.

  18. TOWARD DEVELOPMENT OF A COMMON SOFTWARE APPLICATION PROGRAMMING INTERFACE (API) FOR UNCERTAINTY, SENSITIVITY, AND PARAMETER ESTIMATION METHODS AND TOOLS

    EPA Science Inventory

    The final session of the workshop considered the subject of software technology and how it might be better constructed to support those who develop, evaluate, and apply multimedia environmental models. Two invited presentations were featured along with an extended open discussio...

  19. Effective Process Design for the Production of HIC-Resistant Linepipe Steels

    NASA Astrophysics Data System (ADS)

    Nieto, J.; Elías, T.; López, G.; Campos, G.; López, F.; Garcia, R.; De, Amar K.

    2013-09-01

    Production of slabs for sour service applications requires stringent control in slab internal quality and secondary processing so as to guarantee resistance against hydrogen-induced cracking (HIC). ArcelorMittal Steelmaking facility at Lazaro Cardenas, Mexico had recently implemented key steelmaking and casting processing technologies for production of sound, centerline free slabs for catering to the growing API Linepipe and off-shore market for sour service applications. State-of-the-art steelmaking with use of residual-free Direct-reduced Iron and continuous casting facilities with dynamic soft reduction were introduced for the production of slabs with ultra clean centerline. Introduction of controlled cooling of slabs for atomic hydrogen control well below 2 ppm has enabled production of slabs suitable for excellent HIC-resistant plate processing. Substantial tonnages of slabs were produced for production of API X52-X65 grade plates and pipes for sour service. Stringent quality control at each stage of steelmaking, casting, and slab inspection ensured slabs with excellent internal quality suitable for HIC resistance to be guaranteed in final product (Plates & Pipes). Details of production steps which resulted in successful HIC-resistant slab production have been described in this article.

  20. Developing cloud applications using the e-Science Central platform.

    PubMed

    Hiden, Hugo; Woodman, Simon; Watson, Paul; Cala, Jacek

    2013-01-28

    This paper describes the e-Science Central (e-SC) cloud data processing system and its application to a number of e-Science projects. e-SC provides both software as a service (SaaS) and platform as a service for scientific data management, analysis and collaboration. It is a portable system and can be deployed on both private (e.g. Eucalyptus) and public clouds (Amazon AWS and Microsoft Windows Azure). The SaaS application allows scientists to upload data, edit and run workflows and share results in the cloud, using only a Web browser. It is underpinned by a scalable cloud platform consisting of a set of components designed to support the needs of scientists. The platform is exposed to developers so that they can easily upload their own analysis services into the system and make these available to other users. A representational state transfer-based application programming interface (API) is also provided so that external applications can leverage the platform's functionality, making it easier to build scalable, secure cloud-based applications. This paper describes the design of e-SC, its API and its use in three different case studies: spectral data visualization, medical data capture and analysis, and chemical property prediction.

  1. Developing cloud applications using the e-Science Central platform

    PubMed Central

    Hiden, Hugo; Woodman, Simon; Watson, Paul; Cala, Jacek

    2013-01-01

    This paper describes the e-Science Central (e-SC) cloud data processing system and its application to a number of e-Science projects. e-SC provides both software as a service (SaaS) and platform as a service for scientific data management, analysis and collaboration. It is a portable system and can be deployed on both private (e.g. Eucalyptus) and public clouds (Amazon AWS and Microsoft Windows Azure). The SaaS application allows scientists to upload data, edit and run workflows and share results in the cloud, using only a Web browser. It is underpinned by a scalable cloud platform consisting of a set of components designed to support the needs of scientists. The platform is exposed to developers so that they can easily upload their own analysis services into the system and make these available to other users. A representational state transfer-based application programming interface (API) is also provided so that external applications can leverage the platform's functionality, making it easier to build scalable, secure cloud-based applications. This paper describes the design of e-SC, its API and its use in three different case studies: spectral data visualization, medical data capture and analysis, and chemical property prediction. PMID:23230161

  2. A Web Service Protocol Realizing Interoperable Internet of Things Tasking Capability.

    PubMed

    Huang, Chih-Yuan; Wu, Cheng-Hung

    2016-08-31

    The Internet of Things (IoT) is an infrastructure that interconnects uniquely-identifiable devices using the Internet. By interconnecting everyday appliances, various monitoring, and physical mashup applications can be constructed to improve human's daily life. In general, IoT devices provide two main capabilities: sensing and tasking capabilities. While the sensing capability is similar to the World-Wide Sensor Web, this research focuses on the tasking capability. However, currently, IoT devices created by different manufacturers follow different proprietary protocols and are locked in many closed ecosystems. This heterogeneity issue impedes the interconnection between IoT devices and damages the potential of the IoT. To address this issue, this research aims at proposing an interoperable solution called tasking capability description that allows users to control different IoT devices using a uniform web service interface. This paper demonstrates the contribution of the proposed solution by interconnecting different IoT devices for different applications. In addition, the proposed solution is integrated with the OGC SensorThings API standard, which is a Web service standard defined for the IoT sensing capability. Consequently, the Extended SensorThings API can realize both IoT sensing and tasking capabilities in an integrated and interoperable manner.

  3. WorldWide Telescope in Research and Education

    NASA Astrophysics Data System (ADS)

    Goodman, A.; Fay, J.; Muench, A.; Pepe, A.; Udompraseret, P.; Wong, C.

    2012-09-01

    The WorldWide Telescope computer program, released to researchers and the public as a free resource in 2008 by Microsoft Research, has changed the way the ever-growing Universe of online astronomical data is viewed and understood. The WWT program can be thought of as a scriptable, interactive, richly visual browser of the multi-wavelength Sky as we see it from Earth, and of the Universe as we would travel within it. In its web API format, WWT is being used as a service to display professional research data. In its desktop format, WWT works in concert (thanks to SAMP and other IVOA standards) with more traditional research applications such as ds9, Aladin and TOPCAT. The WWT Ambassadors Program (founded in 2009) recruits and trains astrophysically-literate volunteers (including retirees) who use WWT as a teaching tool in online, classroom, and informal educational settings. Early quantitative studies of WWTA indicate that student experiences with WWT enhance science learning dramatically. Thanks to the wealth of data it can access, and the growing number of services to which it connects, WWT is now a key linking technology in the Seamless Astronomy environment we seek to offer researchers, teachers, and students alike.

  4. Parcels v0.9: prototyping a Lagrangian ocean analysis framework for the petascale age

    NASA Astrophysics Data System (ADS)

    Lange, Michael; van Sebille, Erik

    2017-11-01

    As ocean general circulation models (OGCMs) move into the petascale age, where the output of single simulations exceeds petabytes of storage space, tools to analyse the output of these models will need to scale up too. Lagrangian ocean analysis, where virtual particles are tracked through hydrodynamic fields, is an increasingly popular way to analyse OGCM output, by mapping pathways and connectivity of biotic and abiotic particulates. However, the current software stack of Lagrangian ocean analysis codes is not dynamic enough to cope with the increasing complexity, scale and need for customization of use-cases. Furthermore, most community codes are developed for stand-alone use, making it a nontrivial task to integrate virtual particles at runtime of the OGCM. Here, we introduce the new Parcels code, which was designed from the ground up to be sufficiently scalable to cope with petascale computing. We highlight its API design that combines flexibility and customization with the ability to optimize for HPC workflows, following the paradigm of domain-specific languages. Parcels is primarily written in Python, utilizing the wide range of tools available in the scientific Python ecosystem, while generating low-level C code and using just-in-time compilation for performance-critical computation. We show a worked-out example of its API, and validate the accuracy of the code against seven idealized test cases. This version 0.9 of Parcels is focused on laying out the API, with future work concentrating on support for curvilinear grids, optimization, efficiency and at-runtime coupling with OGCMs.

  5. Effect of Risk Acceptance for Bundled Care Payments on Clinical Outcomes in a High-Volume Total Joint Arthroplasty Practice After Implementation of a Standardized Clinical Pathway.

    PubMed

    Kee, James R; Edwards, Paul K; Barnes, Charles L

    2017-08-01

    The Bundled Payments for Care Improvement (BPCI) initiative and the Arkansas Payment Improvement (API) initiative seek to incentivize reduced costs and improved outcomes compared with the previous fee-for-service model. Before participation, our practice initiated a standardized clinical pathway (CP) to reduce length of stay (LOS), readmissions, and discharge to postacute care facilities. This practice implemented a standardized CP focused on patient education, managing patient expectations, and maximizing cost outcomes. We retrospectively reviewed all primary total joint arthroplasty patients during the initial 2-year "at risk" period for both BPCI and API and determined discharge disposition, LOS, and readmission rate. During the "at risk" period, the average LOS decreased in our total joint arthroplasty patients and our patients discharged home >94%. Patients within the BPCI group had a decreased discharge to home and decreased readmission rates after total hip arthroplasty, but also tended to be older than both API and nonbundled payment patients. While participating in the BPCI and API, continued use of a standardized CP in a high-performing, high-volume total joint practice resulted in maintenance of a low-average LOS. In addition, BPCI patients had similar outcomes after total knee arthroplasty, but had decreased rates of discharge to home and readmission after total hip arthroplasty. Copyright © 2017 Elsevier Inc. All rights reserved.

  6. Search without Boundaries Using Simple APIs

    ERIC Educational Resources Information Center

    Tong, Qi (Helen)

    2009-01-01

    The U.S. Geological Survey (USGS) Library, where the author serves as the digital services librarian, is increasingly challenged to make it easier for users to find information from many heterogeneous information sources. Information is scattered throughout different software applications (i.e., library catalog, federated search engine, link…

  7. ECHO Services: Foundational Middleware for a Science Cyberinfrastructure

    NASA Technical Reports Server (NTRS)

    Burnett, Michael

    2005-01-01

    This viewgraph presentation describes ECHO, an interoperability middleware solution. It uses open, XML-based APIs, and supports net-centric architectures and solutions. ECHO has a set of interoperable registries for both data (metadata) and services, and provides user accounts and a common infrastructure for the registries. It is built upon a layered architecture with extensible infrastructure for supporting community unique protocols. It has been operational since November, 2002 and it available as open source.

  8. A modern Python interface for the Generic Mapping Tools

    NASA Astrophysics Data System (ADS)

    Uieda, L.; Wessel, P.

    2017-12-01

    Figures generated by The Generic Mapping Tools (GMT) are present in countless publications across the Earth sciences. The command-line interface of GMT lends the tool its flexibility but also creates a barrier to entry for begginers. Meanwhile, adoption of the Python programming language has grown across the scientific community. This growth is largely due to the simplicity and low barrier to entry of the language and its ecosystem of tools. Thus, it is not surprising that there have been at least three attempts to create Python interfaces for GMT: gmtpy (github.com/emolch/gmtpy), pygmt (github.com/ian-r-rose/pygmt), and PyGMT (github.com/glimmer-cism/PyGMT). None of these projects are currently active and, with the exception of pygmt, they do not use the GMT Application Programming Interface (API) introduced in GMT 5. The two main Python libraries for plotting data on maps are the matplotlib Basemap toolkit (matplotlib.org/basemap) and Cartopy (scitools.org.uk/cartopy), both of which rely on matplotlib (matplotlib.org) as the backend for generating the figures. Basemap is known to have limitations and is being discontinued. Cartopy is an improvement over Basemap but is still bound by the speed and memory constraints of matplotlib. We present a new Python interface for GMT (GMT/Python) that makes use of the GMT API and of new features being developed for the upcoming GMT 6 release. The GMT/Python library is designed according to the norms and styles of the Python community. The library integrates with the scientific Python ecosystem by using the "virtual files" from the GMT API to implement input and output of Python data types (numpy "ndarray" for tabular data and xarray "Dataset" for grids). Other features include an object-oriented interface for creating figures, the ability to display figures in the Jupyter notebook, and descriptive aliases for GMT arguments (e.g., "region" instead of "R" and "projection" instead of "J"). GMT/Python can also serve as a backend for developing new high-level interfaces, which can help make GMT more accessible to beginners and more intuitive for Python users. GMT/Python is an open-source project hosted on Github (github.com/GenericMappingTools/gmt-python) and is in early stages of development. A first release will accompany the release of GMT 6, which is expected for early 2018.

  9. From WSN towards WoT: Open API Scheme Based on oneM2M Platforms.

    PubMed

    Kim, Jaeho; Choi, Sung-Chan; Ahn, Il-Yeup; Sung, Nak-Myoung; Yun, Jaeseok

    2016-10-06

    Conventional computing systems have been able to be integrated into daily objects and connected to each other due to advances in computing and network technologies, such as wireless sensor networks (WSNs), forming a global network infrastructure, called the Internet of Things (IoT). To support the interconnection and interoperability between heterogeneous IoT systems, the availability of standardized, open application programming interfaces (APIs) is one of the key features of common software platforms for IoT devices, gateways, and servers. In this paper, we present a standardized way of extending previously-existing WSNs towards IoT systems, building the world of the Web of Things (WoT). Based on the oneM2M software platforms developed in the previous project, we introduce a well-designed open API scheme and device-specific thing adaptation software (TAS) enabling WSN elements, such as a wireless sensor node, to be accessed in a standardized way on a global scale. Three pilot services are implemented (i.e., a WiFi-enabled smart flowerpot, voice-based control for ZigBee-connected home appliances, and WiFi-connected AR.Drone control) to demonstrate the practical usability of the open API scheme and TAS modules. Full details on the method of integrating WSN elements into three example systems are described at the programming code level, which is expected to help future researchers in integrating their WSN systems in IoT platforms, such as oneM2M. We hope that the flexibly-deployable, easily-reusable common open API scheme and TAS-based integration method working with the oneM2M platforms will help the conventional WSNs in diverse industries evolve into the emerging WoT solutions.

  10. libChEBI: an API for accessing the ChEBI database.

    PubMed

    Swainston, Neil; Hastings, Janna; Dekker, Adriano; Muthukrishnan, Venkatesh; May, John; Steinbeck, Christoph; Mendes, Pedro

    2016-01-01

    ChEBI is a database and ontology of chemical entities of biological interest. It is widely used as a source of identifiers to facilitate unambiguous reference to chemical entities within biological models, databases, ontologies and literature. ChEBI contains a wealth of chemical data, covering over 46,500 distinct chemical entities, and related data such as chemical formula, charge, molecular mass, structure, synonyms and links to external databases. Furthermore, ChEBI is an ontology, and thus provides meaningful links between chemical entities. Unlike many other resources, ChEBI is fully human-curated, providing a reliable, non-redundant collection of chemical entities and related data. While ChEBI is supported by a web service for programmatic access and a number of download files, it does not have an API library to facilitate the use of ChEBI and its data in cheminformatics software. To provide this missing functionality, libChEBI, a comprehensive API library for accessing ChEBI data, is introduced. libChEBI is available in Java, Python and MATLAB versions from http://github.com/libChEBI, and provides full programmatic access to all data held within the ChEBI database through a simple and documented API. libChEBI is reliant upon the (automated) download and regular update of flat files that are held locally. As such, libChEBI can be embedded in both on- and off-line software applications. libChEBI allows better support of ChEBI and its data in the development of new cheminformatics software. Covering three key programming languages, it allows for the entirety of the ChEBI database to be accessed easily and quickly through a simple API. All code is open access and freely available.

  11. From WSN towards WoT: Open API Scheme Based on oneM2M Platforms

    PubMed Central

    Kim, Jaeho; Choi, Sung-Chan; Ahn, Il-Yeup; Sung, Nak-Myoung; Yun, Jaeseok

    2016-01-01

    Conventional computing systems have been able to be integrated into daily objects and connected to each other due to advances in computing and network technologies, such as wireless sensor networks (WSNs), forming a global network infrastructure, called the Internet of Things (IoT). To support the interconnection and interoperability between heterogeneous IoT systems, the availability of standardized, open application programming interfaces (APIs) is one of the key features of common software platforms for IoT devices, gateways, and servers. In this paper, we present a standardized way of extending previously-existing WSNs towards IoT systems, building the world of the Web of Things (WoT). Based on the oneM2M software platforms developed in the previous project, we introduce a well-designed open API scheme and device-specific thing adaptation software (TAS) enabling WSN elements, such as a wireless sensor node, to be accessed in a standardized way on a global scale. Three pilot services are implemented (i.e., a WiFi-enabled smart flowerpot, voice-based control for ZigBee-connected home appliances, and WiFi-connected AR.Drone control) to demonstrate the practical usability of the open API scheme and TAS modules. Full details on the method of integrating WSN elements into three example systems are described at the programming code level, which is expected to help future researchers in integrating their WSN systems in IoT platforms, such as oneM2M. We hope that the flexibly-deployable, easily-reusable common open API scheme and TAS-based integration method working with the oneM2M platforms will help the conventional WSNs in diverse industries evolve into the emerging WoT solutions. PMID:27782058

  12. Application and API for Real-time Visualization of Ground-motions and Tsunami

    NASA Astrophysics Data System (ADS)

    Aoi, S.; Kunugi, T.; Suzuki, W.; Kubo, T.; Nakamura, H.; Azuma, H.; Fujiwara, H.

    2015-12-01

    Due to the recent progress of seismograph and communication environment, real-time and continuous ground-motion observation becomes technically and economically feasible. K-NET and KiK-net, which are nationwide strong motion networks operated by NIED, cover all Japan by about 1750 stations in total. More than half of the stations transmit the ground-motion indexes and/or waveform data in every second. Traditionally, strong-motion data were recorded by event-triggering based instruments with non-continues telephone line which is connected only after an earthquake. Though the data from such networks mainly contribute to preparations for future earthquakes, huge amount of real-time data from dense network are expected to directly contribute to the mitigation of ongoing earthquake disasters through, e.g., automatic shutdown plants and helping decision-making for initial response. By generating the distribution map of these indexes and uploading them to the website, we implemented the real-time ground motion monitoring system, Kyoshin (strong-motion in Japanese) monitor. This web service (www.kyoshin.bosai.go.jp) started in 2008 and anyone can grasp the current ground motions of Japan. Though this service provides only ground-motion map in GIF format, to take full advantage of real-time strong-motion data to mitigate the ongoing disasters, digital data are important. We have developed a WebAPI to provide real-time data and related information such as ground motions (5 km-mesh) and arrival times estimated from EEW (earthquake early warning). All response data from this WebAPI are in JSON format and are easy to parse. We also developed Kyoshin monitor application for smartphone, 'Kmoni view' using the API. In this application, ground motions estimated from EEW are overlapped on the map with the observed one-second-interval indexes. The application can playback previous earthquakes for demonstration or disaster drill. In mobile environment, data traffic and battery are limited and it is not practical to regularly visualize all the data. The application has automatic starting (pop-up) function triggered by EEW. Similar WebAPI and application for tsunami are being prepared using the pressure data recorded by dense offshore observation network (S-net), which is under construction along the Japan Trench.

  13. Lower HIV prevalence among Asian/Pacific Islander men who have sex with men: a critical review for possible reasons.

    PubMed

    Wei, Chongyi; Raymond, H Fisher; Wong, Frank Y; Silvestre, Anthony J; Friedman, Mark S; Documét, Patricia; McFarland, Willi; Stall, Ron

    2011-04-01

    We conducted a critical literature review for possible reasons that may explain the lower HIV prevalence observed among API MSM compared to MSM of other races/ethnicities. Trends emerging from the literature suggest that traditional individual-level factors-unprotected anal intercourse, substance use, STD prevalence, rates and frequency of HIV testing, and utilization of HIV prevention services-do not appear to be related to the lower HIV prevalence among API MSM. Some evidence suggests that socio-cultural and structural factors might be the more critical forces in determining racial/ethnic disparities of HIV among MSM. For API MSM, these factors include structures of sexual networks, access to and reception of medical care and treatment among HIV-positive MSM, and influences of different levels and types of acculturation. Moreover, emerging risk reduction strategies, such as seroadaptive behaviors, could play a role. Future research should address these factors in intervention design. In addition, better theories of resilience and measurement of strengths and protective factors are needed to enhance the efficacy of HIV interventions.

  14. Unifying Access to National Hydrologic Data Repositories via Web Services

    NASA Astrophysics Data System (ADS)

    Valentine, D. W.; Jennings, B.; Zaslavsky, I.; Maidment, D. R.

    2006-12-01

    The CUAHSI hydrologic information system (HIS) is designed to be a live, multiscale web portal system for accessing, querying, visualizing, and publishing distributed hydrologic observation data and models for any location or region in the United States. The HIS design follows the principles of open service oriented architecture, i.e. system components are represented as web services with well defined standard service APIs. WaterOneFlow web services are the main component of the design. The currently available services have been completely re-written compared to the previous version, and provide programmatic access to USGS NWIS. (steam flow, groundwater and water quality repositories), DAYMET daily observations, NASA MODIS, and Unidata NAM streams, with several additional web service wrappers being added (EPA STORET, NCDC and others.). Different repositories of hydrologic data use different vocabularies, and support different types of query access. Resolving semantic and structural heterogeneities across different hydrologic observation archives and distilling a generic set of service signatures is one of the main scalability challenges in this project, and a requirement in our web service design. To accomplish the uniformity of the web services API, data repositories are modeled following the CUAHSI Observation Data Model. The web service responses are document-based, and use an XML schema to express the semantics in a standard format. Access to station metadata is provided via web service methods, GetSites, GetSiteInfo and GetVariableInfo. The methdods form the foundation of CUAHSI HIS discovery interface and may execute over locally-stored metadata or request the information from remote repositories directly. Observation values are retrieved via a generic GetValues method which is executed against national data repositories. The service is implemented in ASP.Net, and other providers are implementing WaterOneFlow services in java. Reference implementation of WaterOneFlow web services is available. More information about the ongoing development of CUAHSI HIS is available from http://www.cuahsi.org/his/.

  15. Experiences with the Application of Services Oriented Approaches to the Federation of Heterogeneous Geologic Data Resources

    NASA Astrophysics Data System (ADS)

    Cervato, C.; Fils, D.; Bohling, G.; Diver, P.; Greer, D.; Reed, J.; Tang, X.

    2006-12-01

    The federation of databases is not a new endeavor. Great strides have been made e.g. in the health and astrophysics communities. Reviews of those successes indicate that they have been able to leverage off key cross-community core concepts. In its simplest implementation, a federation of databases with identical base schemas that can be extended to address individual efforts, is relatively easy to accomplish. Efforts of groups like the Open Geospatial Consortium have shown methods to geospatially relate data between different sources. We present here a summary of CHRONOS's (http://www.chronos.org) experience with highly heterogeneous data. Our experience with the federation of very diverse databases shows that the wide variety of encoding options for items like locality, time scale, taxon ID, and other key parameters makes it difficult to effectively join data across them. However, the response to this is not to develop one large, monolithic database, which will suffer growth pains due to social, national, and operational issues, but rather to systematically develop the architecture that will enable cross-resource (database, repository, tool, interface) interaction. CHRONOS has accomplished the major hurdle of federating small IT database efforts with service-oriented and XML-based approaches. The application of easy-to-use procedures that allow groups of all sizes to implement and experiment with searches across various databases and to use externally created tools is vital. We are sharing with the geoinformatics community the difficulties with application frameworks, user authentication, standards compliance, and data storage encountered in setting up web sites and portals for various science initiatives (e.g., ANDRILL, EARTHTIME). The ability to incorporate CHRONOS data, services, and tools into the existing framework of a group is crucial to the development of a model that supports and extends the vitality of the small- to medium-sized research effort that is essential for a vibrant scientific community. This presentation will directly address issues of portal development related to JSR-168 and other portal API's as well as issues related to both federated and local directory-based authentication. The application of service-oriented architecture in connection with ReST-based approaches is vital to facilitate service use by experienced and less experienced information technology groups. Application of these services with XML- based schemas allows for the connection to third party tools such a GIS-based tools and software designed to perform a specific scientific analysis. The connection of all these capabilities into a combined framework based on the standard XHTML Document object model and CSS 2.0 standards used in traditional web development will be demonstrated. CHRONOS also utilizes newer client techniques such as AJAX and cross- domain scripting along with traditional server-side database, application, and web servers. The combination of the various components of this architecture creates an environment based on open and free standards that allows for the discovery, retrieval, and integration of tools and data.

  16. Oral biopharmaceutics tools - time for a new initiative - an introduction to the IMI project OrBiTo.

    PubMed

    Lennernäs, H; Aarons, L; Augustijns, P; Beato, S; Bolger, M; Box, K; Brewster, M; Butler, J; Dressman, J; Holm, R; Julia Frank, K; Kendall, R; Langguth, P; Sydor, J; Lindahl, A; McAllister, M; Muenster, U; Müllertz, A; Ojala, K; Pepin, X; Reppas, C; Rostami-Hodjegan, A; Verwei, M; Weitschies, W; Wilson, C; Karlsson, C; Abrahamsson, B

    2014-06-16

    OrBiTo is a new European project within the IMI programme in the area of oral biopharmaceutics tools that includes world leading scientists from nine European universities, one regulatory agency, one non-profit research organization, four SMEs together with scientists from twelve pharmaceutical companies. The OrBiTo project will address key gaps in our knowledge of gastrointestinal (GI) drug absorption and deliver a framework for rational application of predictive biopharmaceutics tools for oral drug delivery. This will be achieved through novel prospective investigations to define new methodologies as well as refinement of existing tools. Extensive validation of novel and existing biopharmaceutics tools will be performed using active pharmaceutical ingredient (API), formulations and supporting datasets from industry partners. A combination of high quality in vitro or in silico characterizations of API and formulations will be integrated into physiologically based in silico biopharmaceutics models capturing the full complexity of GI drug absorption. This approach gives an unparalleled opportunity to initiate a transformational change in industrial research and development to achieve model-based pharmaceutical product development in accordance with the Quality by Design concept. Benefits include an accelerated and more efficient drug candidate selection, formulation development process, particularly for challenging projects such as low solubility molecules (BCS II and IV), enhanced and modified-release formulations, as well as allowing optimization of clinical product performance for patient benefit. In addition, the tools emerging from OrBiTo are expected to significantly reduce demand for animal experiments in the future as well as reducing the number of human bioequivalence studies required to bridge formulations after manufacturing or composition changes. Copyright © 2013 Elsevier B.V. All rights reserved.

  17. Quantitative Component Analysis of Solid Mixtures by Analyzing Time Domain 1H and 19F T1 Saturation Recovery Curves (qSRC).

    PubMed

    Stueber, Dirk; Jehle, Stefan

    2017-07-01

    Prevalent polymorphism and complicated phase behavior of active pharmaceutical ingredients (APIs) often result in remarkable differences in the respective biochemical and physical API properties. Consequently, API form characterization and quantification play a central role in the pharmaceutical industry from early drug development to manufacturing. Here we present a novel and proficient quantification protocol for solid mixtures (qSRC) based on the measurement and mathematical fitting of T 1 nuclear magnetic resonance (NMR) saturation recovery curves collected on a bench top time-domain NMR instrument. The saturation recovery curves of the relevant pure components are used as fingerprints. Employing a bench top NMR instrument possesses clear benefits. These instruments exhibit a small footprint, do not present any special requirements on lab space, and required sample handling is simple and fast. The qSRC analysis can easily be conducted in a conventional laboratory setting as well as in an industrial production environment, making it a versatile tool with the potential for widespread application. The accuracy and efficiency of the qSRC method is illustrated using 1 H and 19 F T 1 data of selected pharmaceutical model compounds, as well as utilizing 1 H T 1 data of an actual binary API anhydrous polymorph system of a Merck & Co., Inc. compound formerly developed as a hepatitis C virus drug. Copyright © 2017 American Pharmacists Association®. Published by Elsevier Inc. All rights reserved.

  18. Using Enabling Technologies to Facilitate the Comparison of Satellite Observations with the Model Forecasts for Hurricane Study

    NASA Astrophysics Data System (ADS)

    Li, P.; Knosp, B.; Hristova-Veleva, S. M.; Niamsuwan, N.; Johnson, M. P.; Shen, T. P. J.; Tanelli, S.; Turk, J.; Vu, Q. A.

    2014-12-01

    Due to their complexity and volume, the satellite data are underutilized in today's hurricane research and operations. To better utilize these data, we developed the JPL Tropical Cyclone Information System (TCIS) - an Interactive Data Portal providing fusion between Near-Real-Time satellite observations and model forecasts to facilitate model evaluation and improvement. We have collected satellite observations and model forecasts in the Atlantic Basin and the East Pacific for the hurricane seasons since 2010 and supported the NASA Airborne Campaigns for Hurricane Study such as the Genesis and Rapid Intensification Processes (GRIP) in 2010 and the Hurricane and Severe Storm Sentinel (HS3) from 2012 to 2014. To enable the direct inter-comparisons of the satellite observations and the model forecasts, the TCIS was integrated with the NASA Earth Observing System Simulator Suite (NEOS3) to produce synthetic observations (e.g. simulated passive microwave brightness temperatures) from a number of operational hurricane forecast models (HWRF and GFS). An automated process was developed to trigger NEOS3 simulations via web services given the location and time of satellite observations, monitor the progress of the NEOS3 simulations, display the synthetic observation and ingest them into the TCIS database when they are done. In addition, three analysis tools, the joint PDF analysis of the brightness temperatures, ARCHER for finding the storm-center and the storm organization and the Wave Number Analysis tool for storm asymmetry and morphology analysis were integrated into TCIS to provide statistical and structural analysis on both observed and synthetic data. Interactive tools were built in the TCIS visualization system to allow the spatial and temporal selections of the datasets, the invocation of the tools with user specified parameters, and the display and the delivery of the results. In this presentation, we will describe the key enabling technologies behind the design of the TCIS interactive data portal and analysis tools, including the spatial database technology for the representation and query of the level 2 satellite data, the automatic process flow using web services, the interactive user interface using the Google Earth API, and a common and expandable Python wrapper to invoke the analysis tools.

  19. Accompanying Partners of International Students: Reflections on Three Issues

    ERIC Educational Resources Information Center

    Cui, Dan; Arthur, Nancy; Domene, José F.

    2017-01-01

    This paper reviews and critiques the existing literature on accompanying partners of international students (APIS), who are often an ignored population in programs and services for the internationalization of Canadian higher education. Particularly, we identify three issues. First, we argue that current research on this group overwhelmingly…

  20. Towards easing the configuration and new team member accommodation for open source software based portals

    NASA Astrophysics Data System (ADS)

    Fu, L.; West, P.; Zednik, S.; Fox, P. A.

    2013-12-01

    For simple portals such as vocabulary based services, which contain small amounts of data and require only hyper-textual representation, it is often an overkill to adopt the whole software stack of database, middleware and front end, or to use a general Web development framework as the starting point of development. Directly combining open source software is a much more favorable approach. However, our experience with the Coastal and Marine Spatial Planning Vocabulary (CMSPV) service portal shows that there are still issues such as system configuration and accommodating a new team member that need to be handled carefully. In this contribution, we share our experience in the context of the CMSPV portal, and focus on the tools and mechanisms we've developed to ease the configuration job and the incorporation process of new project members. We discuss the configuration issues that arise when we don't have complete control over how the software in use is configured and need to follow existing configuration styles that may not be well documented, especially when multiple pieces of such software need to work together as a combined system. As for the CMSPV portal, it is built on two pieces of open source software that are still under rapid development: a Fuseki data server and Epimorphics Linked Data API (ELDA) front end. Both lack mature documentation and tutorials. We developed comparison and labeling tools to ease the problem of system configuration. Another problem that slowed down the project is that project members came and went during the development process, so new members needed to start with a partially configured system and incomplete documentation left by old members. We developed documentation/tutorial maintenance mechanisms based on our comparison and labeling tools to make it easier for the new members to be incorporated into the project. These tools and mechanisms also provided benefit to other projects that reused the software components from the CMSPV system.

  1. panMetaDocs, eSciDoc, and DOIDB - an infrastructure for the curation and publication of file-based datasets for 'GFZ Data Services'

    NASA Astrophysics Data System (ADS)

    Ulbricht, Damian; Elger, Kirsten; Bertelmann, Roland; Klump, Jens

    2016-04-01

    With the foundation of DataCite in 2009 and the technical infrastructure installed in the last six years it has become very easy to create citable dataset DOIs. Nowadays, dataset DOIs are increasingly accepted and required by journals in reference lists of manuscripts. In addition, DataCite provides usage statistics [1] of assigned DOIs and offers a public search API to make research data count. By linking related information to the data, they become more useful for future generations of scientists. For this purpose, several identifier systems, as ISBN for books, ISSN for journals, DOI for articles or related data, Orcid for authors, and IGSN for physical samples can be attached to DOIs using the DataCite metadata schema [2]. While these are good preconditions to publish data, free and open solutions that help with the curation of data, the publication of research data, and the assignment of DOIs in one software seem to be rare. At GFZ Potsdam we built a modular software stack that is made of several free and open software solutions and we established 'GFZ Data Services'. 'GFZ Data Services' provides storage, a metadata editor for publication and a facility to moderate minted DOIs. All software solutions are connected through web APIs, which makes it possible to reuse and integrate established software. Core component of 'GFZ Data Services' is an eSciDoc [3] middleware that is used as central storage, and has been designed along the OAIS reference model for digital preservation. Thus, data are stored in self-contained packages that are made of binary file-based data and XML-based metadata. The eSciDoc infrastructure provides access control to data and it is able to handle half-open datasets, which is useful in embargo situations when a subset of the research data are released after an adequate period. The data exchange platform panMetaDocs [4] makes use of eSciDoc's REST API to upload file-based data into eSciDoc and uses a metadata editor [5] to annotate the files with metadata. The metadata editor has a user-friendly interface with nominal lists, extensive explanations, and an interactive mapping tool to provide assistance to scientists describing the data. It is possible to deposit metadata templates to fill certain fields with default values. The metadata editor generates metadata in the schemas ISO19139, NASA GCMD DIF, and DataCite and could be extended for other schemas. panMetaDocs is able to mint dataset DOIs through DOIDB, which is our component to moderate dataset DOIs issued through 'GFZ Data Services'. DOIDB accepts metadata in the schemas ISO19139, DIF, and DataCite. In addition, DOIDB provides an OAI-PMH interface to disseminate all deposited metadata to data portals. The presentation of datasets on DOI landing pages is done though XSLT stylesheet transformation of the XML-based metadata. The landing pages have been designed to meet needs of scientists. We are able to render the metadata to different layouts. Furthermore, additional information about datasets and publications is assembled into the webpage by querying public databases on the internet. The work presented here will focus on technical details of the software stack. [1] http://stats.datacite.org [2] http://www.dlib.org/dlib/january11/starr/01starr.html [3] http://www.escidoc.org [4] http://panmetadocs.sf.net [5] http://github.com/ulbricht

  2. Unleashing Geophysics Data with Modern Formats and Services

    NASA Astrophysics Data System (ADS)

    Ip, Alex; Brodie, Ross C.; Druken, Kelsey; Bastrakova, Irina; Evans, Ben; Kemp, Carina; Richardson, Murray; Trenham, Claire; Wang, Jingbo; Wyborn, Lesley

    2016-04-01

    Geoscience Australia (GA) is the national steward of large volumes of geophysical data extending over the entire Australasian region and spanning many decades. The volume and variety of data which must be managed, coupled with the increasing need to support machine-to-machine data access, mean that the old "click-and-ship" model delivering data as downloadable files for local analysis is rapidly becoming unviable - a "big data" problem not unique to geophysics. The Australian Government, through the Research Data Services (RDS) Project, recently funded the Australian National Computational Infrastructure (NCI) to organize a wide range of Earth Systems data from diverse collections including geoscience, geophysics, environment, climate, weather, and water resources onto a single High Performance Data (HPD) Node. This platform, which now contains over 10 petabytes of data, is called the National Environmental Research Data Interoperability Platform (NERDIP), and is designed to facilitate broad user access, maximise reuse, and enable integration. GA has contributed several hundred terabytes of geophysical data to the NERDIP. Historically, geophysical datasets have been stored in a range of formats, with metadata of varying quality and accessibility, and without standardised vocabularies. This has made it extremely difficult to aggregate original data from multiple surveys (particularly un-gridded geophysics point/line data) into standard formats suited to High Performance Computing (HPC) environments. To address this, it was decided to use the NERDIP-preferred Hierarchical Data Format (HDF) 5, which is a proven, standard, open, self-describing and high-performance format supported by extensive software tools, libraries and data services. The Network Common Data Form (NetCDF) 4 API facilitates the use of data in HDF5, whilst the NetCDF Climate & Forecasting conventions (NetCDF-CF) further constrain NetCDF4/HDF5 data so as to provide greater inherent interoperability. The first geophysical data collection selected for transformation by GA was Airborne ElectroMagnetics (AEM) data which was held in proprietary-format files, with associated ISO 19115 metadata held in a separate relational database. Existing NetCDF-CF metadata profiles were enhanced to cover AEM and other geophysical data types, and work is underway to formalise the new geophysics vocabulary as a proposed extension to the Climate & Forecasting conventions. The richness and flexibility of HDF5's internal indexing mechanisms has allowed lossless restructuring of the AEM data for efficient storage, subsetting and access via either the NetCDF4/HDF5 APIs or Open-source Project for a Network Data Access Protocol (OPeNDAP) data services. This approach not only supports large-scale HPC processing, but also interactive access to a wide range of geophysical data in user-friendly environments such as iPython notebooks and more sophisticated cloud-enabled portals such as the Virtual Geophysics Laboratory (VGL). As multidimensional AEM datasets are relatively complex compared to other geophysical data types, the general approach employed in this project for modernizing AEM data is likely to be applicable to other geophysics data types. When combined with the use of standards-based data services and APIs, a coordinated, systematic modernisation will result in vastly improved accessibility to, and usability of, geophysical data in a wide range of computational environments both within and beyond the geophysics community.

  3. Current Standardization and Cooperative Efforts Related to Industrial Information Infrastructures.

    DTIC Science & Technology

    1993-05-01

    Data Management Systems: Components used to store, manage, and retrieve data. Data management includes knowledge bases, database management...Application Development Tools and Methods X/Open and POSIX APIs Integrated Design Support System (IDS) Knowledge -Based Systems (KBS) Application...IDEFlx) Yourdon Jackson System Design (JSD) Knowledge -Based Systems (KBSs) Structured Systems Development (SSD) Semantic Unification Meta-Model

  4. Outside the Framework of Thinkable Thought: The Modern Orchestration Project

    ERIC Educational Resources Information Center

    Gattegno, Eliot Aron

    2010-01-01

    In today's world of too much information, context--not content--is king. This proposal is for the development of an unparalleled sonic analysis tool that converts audio files into musical score notation and a Web site (API) to collect manage and preserve information about the musical sounds analyzed, as well as music scores, videos, and articles…

  5. Spatial accessibility to healthcare services in Shenzhen, China: improving the multi-modal two-step floating catchment area method by estimating travel time via online map APIs.

    PubMed

    Tao, Zhuolin; Yao, Zaoxing; Kong, Hui; Duan, Fei; Li, Guicai

    2018-05-09

    Shenzhen has rapidly grown into a megacity in the recent decades. It is a challenging task for the Shenzhen government to provide sufficient healthcare services. The spatial configuration of healthcare services can influence the convenience for the consumers to obtain healthcare services. Spatial accessibility has been widely adopted as a scientific measurement for evaluating the rationality of the spatial configuration of healthcare services. The multi-modal two-step floating catchment area (2SFCA) method is an important advance in the field of healthcare accessibility modelling, which enables the simultaneous assessment of spatial accessibility via multiple transport modes. This study further develops the multi-modal 2SFCA method by introducing online map APIs to improve the estimation of travel time by public transit or by car respectively. As the results show, the distribution of healthcare accessibility by multi-modal 2SFCA shows significant spatial disparity. Moreover, by dividing the multi-modal accessibility into car-mode and transit-mode accessibility, this study discovers that the transit-mode subgroup is disadvantaged in the competition for healthcare services with the car-mode subgroup. The disparity in transit-mode accessibility is the main reason of the uneven pattern of healthcare accessibility in Shenzhen. The findings suggest improving the public transit conditions for accessing healthcare services to reduce the disparity of healthcare accessibility. More healthcare services should be allocated in the eastern and western Shenzhen, especially sub-districts in Dapeng District and western Bao'an District. As these findings cannot be drawn by the traditional single-modal 2SFCA method, the advantage of the multi-modal 2SFCA method is significant to both healthcare studies and healthcare system planning.

  6. Helioviewer.org: Simple Solar and Heliospheric Data Visualization

    NASA Astrophysics Data System (ADS)

    Hughitt, V. K.; Ireland, J.; Mueller, D.

    2011-12-01

    Helioviewer.org is a free and open-source web application for exploring solar physics data in a simple and intuitive manner. Over the past several years, Helioviewer.org has enabled thousands of users from across the globe to explore the inner heliosphere, providing access to over ten million images from the SOHO, SDO, and STEREO missions. While Helioviewer.org has seen a surge in use by the public in recent months, it is still ultimately a science tool. The newest version of Helioviewer.org provides access to science-quality data for all available images through the Virtual Solar Observatory (VSO). In addition to providing a powerful platform for browsing heterogeneous sets of solar data, Helioviewer.org also seeks to be as flexible and extensible as possible, providing access to much of its functionality via a simple Application Programming Interface (API). Recently, the Helioviewer.org API was used for two such applications: a Wordpress plugin, and a Python library for solar physics data analysis (SunPy). These applications are discussed and examples of API usage are provided. Finally, Helioviewer.org is undergoing continual development, with new features being added on a regular basis. Recent updates to Helioviewer.org are discussed, along with a preview of things to come.

  7. An integrated approach for prioritizing pharmaceuticals found in the environment for risk assessment, monitoring and advanced research.

    PubMed

    Caldwell, Daniel J; Mastrocco, Frank; Margiotta-Casaluci, Luigi; Brooks, Bryan W

    2014-11-01

    Numerous active pharmaceutical ingredients (APIs), approved prior to enactment of detailed environmental risk assessment (ERA) guidance in the EU in 2006, have been detected in surface waters as a result of advancements in analytical technologies. Without adequate knowledge of the potential hazards these APIs may pose, assessing their environmental risk is challenging. As it would be impractical to commence hazard characterization and ERA en masse, several approaches to prioritizing substances for further attention have been published. Here, through the combination of three presentations given at a recent conference, "Pharmaceuticals in the Environment, Is there a problem?" (Nîmes, France, June 2013) we review several of these approaches, identify salient components, and present available techniques and tools that could facilitate a pragmatic, scientifically sound approach to prioritizing APIs for advanced study or ERA and, where warranted, fill critical data gaps through targeted, intelligent testing. We further present a modest proposal to facilitate future prioritization efforts and advanced research studies that incorporates mammalian pharmacology data (e.g., adverse outcomes pathways and the fish plasma model) and modeled exposure data based on pharmaceutical use. Copyright © 2014 Elsevier Ltd. All rights reserved.

  8. Hot-stage microscopy for determination of API fragmentation: comparison with other methods.

    PubMed

    Šimek, Michal; Grünwaldová, Veronika; Kratochvíl, Bohumil

    2016-08-01

    Although the fragmentation of the active pharmaceutical ingredient (API) is a phenomenon that is mentioned in many literature sources, no well-suited analytical tools for its investigation are currently known. We used the hot-stage microscopy method, already presented in our previous work, and studied the real fragmentation of the tadalafil particles in model tablets which were prepared under different compaction pressures. The morphology, spectral imaging and evaluation of plastic and elastic energies were also analyzed to support the hot-stage method. The prepared blend of tadalafil and excipients was compacted under a several forces from 5 to 35 kN to reveal the trend of fragmentation. The exact fragmentation of tadalafil with increased compaction pressure was revealed by the hot-stage microscopic method and it was in good agreement with plastic and elastic energies. Conversely, spectral imaging, which is being used for this analysis, was considered to be inaccurate methodology as mainly agglomerates, not individual particles, were measured. The availability of the hot-stage microscopic method equips pharmaceutical scientists with an in vitro assessment technique that will more reliably determine the fragmentation of the API in finished tablets and the behavior of the particles when compacted.

  9. Active pharmaceutical ingredient (API) production involving continuous processes--a process system engineering (PSE)-assisted design framework.

    PubMed

    Cervera-Padrell, Albert E; Skovby, Tommy; Kiil, Søren; Gani, Rafiqul; Gernaey, Krist V

    2012-10-01

    A systematic framework is proposed for the design of continuous pharmaceutical manufacturing processes. Specifically, the design framework focuses on organic chemistry based, active pharmaceutical ingredient (API) synthetic processes, but could potentially be extended to biocatalytic and fermentation-based products. The method exploits the synergic combination of continuous flow technologies (e.g., microfluidic techniques) and process systems engineering (PSE) methods and tools for faster process design and increased process understanding throughout the whole drug product and process development cycle. The design framework structures the many different and challenging design problems (e.g., solvent selection, reactor design, and design of separation and purification operations), driving the user from the initial drug discovery steps--where process knowledge is very limited--toward the detailed design and analysis. Examples from the literature of PSE methods and tools applied to pharmaceutical process design and novel pharmaceutical production technologies are provided along the text, assisting in the accumulation and interpretation of process knowledge. Different criteria are suggested for the selection of batch and continuous processes so that the whole design results in low capital and operational costs as well as low environmental footprint. The design framework has been applied to the retrofit of an existing batch-wise process used by H. Lundbeck A/S to produce an API: zuclopenthixol. Some of its batch operations were successfully converted into continuous mode, obtaining higher yields that allowed a significant simplification of the whole process. The material and environmental footprint of the process--evaluated through the process mass intensity index, that is, kg of material used per kg of product--was reduced to half of its initial value, with potential for further reduction. The case-study includes reaction steps typically used by the pharmaceutical industry featuring different characteristic reaction times, as well as L-L separation and distillation-based solvent exchange steps, and thus constitutes a good example of how the design framework can be useful to efficiently design novel or already existing API manufacturing processes taking advantage of continuous processes. Copyright © 2012 Elsevier B.V. All rights reserved.

  10. The tissue microarray data exchange specification: A document type definition to validate and enhance XML data

    PubMed Central

    Nohle, David G; Ayers, Leona W

    2005-01-01

    Background The Association for Pathology Informatics (API) Extensible Mark-up Language (XML) TMA Data Exchange Specification (TMA DES) proposed in April 2003 provides a community-based, open source tool for sharing tissue microarray (TMA) data in a common format. Each tissue core within an array has separate data including digital images; therefore an organized, common approach to produce, navigate and publish such data facilitates viewing, sharing and merging TMA data from different laboratories. The AIDS and Cancer Specimen Resource (ACSR) is a HIV/AIDS tissue bank consortium sponsored by the National Cancer Institute (NCI) Division of Cancer Treatment and Diagnosis (DCTD). The ACSR offers HIV-related malignancies and uninfected control tissues in microarrays (TMA) accompanied by de-identified clinical data to approved researchers. Exporting our TMA data into the proposed API specified format offers an opportunity to evaluate the API specification in an applied setting and to explore its usefulness. Results A document type definition (DTD) that governs the allowed common data elements (CDE) in TMA DES export XML files was written, tested and evolved and is in routine use by the ACSR. This DTD defines TMA DES CDEs which are implemented in an external file that can be supplemented by internal DTD extensions for locally defined TMA data elements (LDE). Conclusion ACSR implementation of the TMA DES demonstrated the utility of the specification and allowed application of a DTD to validate the language of the API specified XML elements and to identify possible enhancements within our TMA data management application. Improvements to the specification have additionally been suggested by our experience in importing other institution's exported TMA data. Enhancements to TMA DES to remove ambiguous situations and clarify the data should be considered. Better specified identifiers and hierarchical relationships will make automatic use of the data possible. Our tool can be used to reorder data and add identifiers; upgrading data for changes in the specification can be automatically accomplished. Using a DTD (optionally reflecting our proposed enhancements) can provide stronger validation of exported TMA data. PMID:15871741

  11. Applying standards to ICT models, tools and data in Europe to improve river basin networks and spread innovation on water sector

    NASA Astrophysics Data System (ADS)

    Pesquer, Lluís; Jirka, Simon; van de Giesen, Nick; Masó, Joan; Stasch, Christoph; Van Nooyen, Ronald; Prat, Ester; Pons, Xavier

    2015-04-01

    This work describes the strategy of the European Horizon 2020 project WaterInnEU. Its vision is to enhance the exploitation of EU funded ICT models, tools, protocols and policy briefs related to the water sector and to establish suitable conditions for new market opportunities based on these offerings. The main goals are: • Connect the research results and developments of previous EU funded activities with the already existing data available on European level and also with to the companies that are able to offer products and services based on these tools and data. • Offer an independent marketplace platform complemented by technical and commercial expertise as a service for users to allow the access to products and services best fitting their priorities, capabilities and procurement processes. One of the pillars of WaterInnEU is to stimulate and prioritize the application of international standards into ICT tools and policy briefs. The standardization of formats, services and processes will allow for a harmonized water management between different sectors, fragmented areas and scales (local, regional or international) approaches. Several levels of interoperability will be addressed: • Syntactic: Connecting system and tools together: Syntactic interoperability allows for client and service tools to automatically discover, access, and process data and information (query and exchange parts of a database) and to connect each other in process chains. The discovery of water related data is achieved using metadata cataloguing standards and, in particular, the one adopted by the INSPIRE directive: OGC Catalogue Service for the Web (CSW). • Semantic: Sharing a pan-European conceptual framework This is the ability of computer systems to exchange data with unambiguous, shared meaning. The project therefore addresses not only the packaging of data (syntax), but also the simultaneous transmission of the meaning with the data (semantics). This is accomplished by linking each data element to a controlled, shared vocabulary. In Europe, INSPIRE defines a shared vocabulary and its associated links to an ontology. For hydrographical information this can be used as a baseline. • Organizational: Harmonizing policy aspects This level of interoperability deals with operational methodologies and procedures that organizations use to administrate their own data and processing capabilities and to share those capabilities with others. This layer is addressed by the adoption of common policy briefs that facilitate both robust protocols and flexibility to interact with others. • Data visualization: Making data easy to see The WMS and WMTS standards are the most commonly used geographic information visualization standards for sharing information in web portals. Our solution will incorporate a quality extension of these standards for visualizing data quality as nested layers linked to the different data sets. In the presented approach, the use of standards can be seen twofold: the tools and products should leverage standards wherever possible to ensure interoperability between solution providers, and the platform itself must utilize standards as much as possible, to allow for example the integration with other systems through open APIs or the description of available items.

  12. Towards a single seismological service infrastructure in Europe

    NASA Astrophysics Data System (ADS)

    Spinuso, A.; Trani, L.; Frobert, L.; Van Eck, T.

    2012-04-01

    In the last five year services and data providers, within the seismological community in Europe, focused their efforts in migrating the way of opening their archives towards a Service Oriented Architecture (SOA). This process tries to follow pragmatically the technological trends and available solutions aiming at effectively improving all the data stewardship activities. These advancements are possible thanks to the cooperation and the follow-ups of several EC infrastructural projects that, by looking at general purpose techniques, combine their developments envisioning a multidisciplinary platform for the earth observation as the final common objective (EPOS, Earth Plate Observation System) One of the first results of this effort is the Earthquake Data Portal (http://www.seismicportal.eu), which provides a collection of tools to discover, visualize and access a variety of seismological data sets like seismic waveform, accelerometric data, earthquake catalogs and parameters. The Portal offers a cohesive distributed search environment, linking data search and access across multiple data providers through interactive web-services, map-based tools and diverse command-line clients. Our work continues under other EU FP7 projects. Here we will address initiatives in two of those projects. The NERA, (Network of European Research Infrastructures for Earthquake Risk Assessment and Mitigation) project will implement a Common Services Architecture based on OGC services APIs, in order to provide Resource-Oriented common interfaces across the data access and processing services. This will improve interoperability between tools and across projects, enabling the development of higher-level applications that can uniformly access the data and processing services of all participants. This effort will be conducted jointly with the VERCE project (Virtual Earthquake and Seismology Research Community for Europe). VERCE aims to enable seismologists to exploit the wealth of seismic data within a data-intensive computation framework, which will be tailored to the specific needs of the community. It will provide a new interoperable infrastructure, as the computational backbone laying behind the publicly available interfaces. VERCE will have to face the challenges of implementing a service oriented architecture providing an efficient layer between the Data and the Grid infrastructures, coupling HPC data analysis and HPC data modeling applications through the execution of workflows and data sharing mechanism. Online registries of interoperable worklflow components, storage of intermediate results and data provenance are those aspects that are currently under investigations to make the VERCE facilities usable from a large scale of users, data and service providers. For such purposes the adoption of a Digital Object Architecture, to create online catalogs referencing and describing semantically all these distributed resources, such as datasets, computational processes and derivative products, is seen as one of the viable solution to monitor and steer the usage of the infrastructure, increasing its efficiency and the cooperation among the community.

  13. Framework for End-User Programming of Cross-Smart Space Applications

    PubMed Central

    Palviainen, Marko; Kuusijärvi, Jarkko; Ovaska, Eila

    2012-01-01

    Cross-smart space applications are specific types of software services that enable users to share information, monitor the physical and logical surroundings and control it in a way that is meaningful for the user's situation. For developing cross-smart space applications, this paper makes two main contributions: it introduces (i) a component design and scripting method for end-user programming of cross-smart space applications and (ii) a backend framework of components that interwork to support the brunt of the RDFScript translation, and the use and execution of ontology models. Before end-user programming activities, the software professionals must develop easy-to-apply Driver components for the APIs of existing software systems. Thereafter, end-users are able to create applications from the commands of the Driver components with the help of the provided toolset. The paper also introduces the reference implementation of the framework, tools for the Driver component development and end-user programming of cross-smart space applications and the first evaluation results on their application. PMID:23202169

  14. Globus platform-as-a-service for collaborative science applications

    DOE PAGES

    Ananthakrishnan, Rachana; Chard, Kyle; Foster, Ian; ...

    2014-03-13

    Globus, developed as Software-as-a-Service (SaaS) for research data management, also provides APIs that constitute a flexible and powerful Platform-as-a-Service (PaaS) to which developers can outsource data management activities such as transfer and sharing, as well as identity, profile and group management. By providing these frequently important but always challenging capabilities as a service, accessible over the network, Globus PaaS streamlines web application development and makes it easy for individuals, teams, and institutions to create collaborative applications such as science gateways for science communities. We introduce the capabilities of this platform and review representative applications.

  15. Globus Platform-as-a-Service for Collaborative Science Applications.

    PubMed

    Ananthakrishnan, Rachana; Chard, Kyle; Foster, Ian; Tuecke, Steven

    2015-02-01

    Globus, developed as Software-as-a-Service (SaaS) for research data management, also provides APIs that constitute a flexible and powerful Platform-as-a-Service (PaaS) to which developers can outsource data management activities such as transfer and sharing, as well as identity, profile and group management. By providing these frequently important but always challenging capabilities as a service, accessible over the network, Globus PaaS streamlines web application development and makes it easy for individuals, teams, and institutions to create collaborative applications such as science gateways for science communities. We introduce the capabilities of this platform and review representative applications.

  16. Globus Platform-as-a-Service for Collaborative Science Applications

    PubMed Central

    Ananthakrishnan, Rachana; Chard, Kyle; Foster, Ian; Tuecke, Steven

    2014-01-01

    Globus, developed as Software-as-a-Service (SaaS) for research data management, also provides APIs that constitute a flexible and powerful Platform-as-a-Service (PaaS) to which developers can outsource data management activities such as transfer and sharing, as well as identity, profile and group management. By providing these frequently important but always challenging capabilities as a service, accessible over the network, Globus PaaS streamlines web application development and makes it easy for individuals, teams, and institutions to create collaborative applications such as science gateways for science communities. We introduce the capabilities of this platform and review representative applications. PMID:25642152

  17. The E3 ubiquitin ligases β-TrCP and FBXW7 cooperatively mediates GSK3-dependent Mcl-1 degradation induced by the Akt inhibitor API-1, resulting in apoptosis.

    PubMed

    Ren, Hui; Koo, Junghui; Guan, Baoxiang; Yue, Ping; Deng, Xingming; Chen, Mingwei; Khuri, Fadlo R; Sun, Shi-Yong

    2013-11-22

    The novel Akt inhibitor, API-1, induces apoptosis through undefined mechanisms. The current study focuses on revealing the mechanisms by which API-1 induces apoptosis. API-1 rapidly and potently reduced the levels of Mcl-1 primarily in API-1-senstive lung cancer cell lines. Ectopic expression of Mcl-1 protected cells from induction of apoptosis by API-1. API-1 treatment decreased the half-life of Mcl-1, whereas inhibition of the proteasome with MG132 rescued Mcl-1 reduction induced by API-1. API-1 decreased Mcl-1 levels accompanied with a rapid increase in Mcl-1 phosphorylation (S159/T163). Moreover, inhibition of GSK3 inhibited Mcl-1 phosphorylation and reduction induced by API-1 and antagonized the effect of API-1 on induction of apoptosis. Knockdown of either FBXW7 or β-TrCP alone, both of which are E3 ubiquitin ligases involved in Mcl-1 degradation, only partially rescued Mcl-1 reduction induced by API-1. However, double knockdown of both E3 ubiquitin ligases enhanced the rescue of API-1-induced Mcl-1 reduction. API-1 induces GSK3-dependent, β-TrCP- and FBXW7-mediated Mcl-1 degradation, resulting in induction of apoptosis.

  18. The E3 ubiquitin ligases β-TrCP and FBXW7 cooperatively mediates GSK3-dependent Mcl-1 degradation induced by the Akt inhibitor API-1, resulting in apoptosis

    PubMed Central

    2013-01-01

    Background The novel Akt inhibitor, API-1, induces apoptosis through undefined mechanisms. The current study focuses on revealing the mechanisms by which API-1 induces apoptosis. Results API-1 rapidly and potently reduced the levels of Mcl-1 primarily in API-1-senstive lung cancer cell lines. Ectopic expression of Mcl-1 protected cells from induction of apoptosis by API-1. API-1 treatment decreased the half-life of Mcl-1, whereas inhibition of the proteasome with MG132 rescued Mcl-1 reduction induced by API-1. API-1 decreased Mcl-1 levels accompanied with a rapid increase in Mcl-1 phosphorylation (S159/T163). Moreover, inhibition of GSK3 inhibited Mcl-1 phosphorylation and reduction induced by API-1 and antagonized the effect of API-1 on induction of apoptosis. Knockdown of either FBXW7 or β-TrCP alone, both of which are E3 ubiquitin ligases involved in Mcl-1 degradation, only partially rescued Mcl-1 reduction induced by API-1. However, double knockdown of both E3 ubiquitin ligases enhanced the rescue of API-1-induced Mcl-1 reduction. Conclusions API-1 induces GSK3-dependent, β-TrCP- and FBXW7-mediated Mcl-1 degradation, resulting in induction of apoptosis. PMID:24261825

  19. Genome-wide analysis of signatures of selection in populations of African honey bees (Apis mellifera) using new web-based tools.

    PubMed

    Fuller, Zachary L; Niño, Elina L; Patch, Harland M; Bedoya-Reina, Oscar C; Baumgarten, Tracey; Muli, Elliud; Mumoki, Fiona; Ratan, Aakrosh; McGraw, John; Frazier, Maryann; Masiga, Daniel; Schuster, Stephen; Grozinger, Christina M; Miller, Webb

    2015-07-10

    With the development of inexpensive, high-throughput sequencing technologies, it has become feasible to examine questions related to population genetics and molecular evolution of non-model species in their ecological contexts on a genome-wide scale. Here, we employed a newly developed suite of integrated, web-based programs to examine population dynamics and signatures of selection across the genome using several well-established tests, including F ST, pN/pS, and McDonald-Kreitman. We applied these techniques to study populations of honey bees (Apis mellifera) in East Africa. In Kenya, there are several described A. mellifera subspecies, which are thought to be localized to distinct ecological regions. We performed whole genome sequencing of 11 worker honey bees from apiaries distributed throughout Kenya and identified 3.6 million putative single-nucleotide polymorphisms. The dense coverage allowed us to apply several computational procedures to study population structure and the evolutionary relationships among the populations, and to detect signs of adaptive evolution across the genome. While there is considerable gene flow among the sampled populations, there are clear distinctions between populations from the northern desert region and those from the temperate, savannah region. We identified several genes showing population genetic patterns consistent with positive selection within African bee populations, and between these populations and European A. mellifera or Asian Apis florea. These results lay the groundwork for future studies of adaptive ecological evolution in honey bees, and demonstrate the use of new, freely available web-based tools and workflows ( http://usegalaxy.org/r/kenyanbee ) that can be applied to any model system with genomic information.

  20. OpenAQ: A Platform to Aggregate and Freely Share Global Air Quality Data

    NASA Astrophysics Data System (ADS)

    Hasenkopf, C. A.; Flasher, J. C.; Veerman, O.; DeWitt, H. L.

    2015-12-01

    Thousands of ground-based air quality monitors around the world publicly publish real-time air quality data; however, researchers and the public do not have access to this information in the ways most useful to them. Often, air quality data are posted on obscure websites showing only current values, are programmatically inaccessible, and/or are in inconsistent data formats across sites. Yet, historical and programmatic access to such a global dataset would be transformative to several scientific fields, from epidemiology to low-cost sensor technologies to estimates of ground-level aerosol by satellite retrievals. To increase accessibility and standardize this disparate dataset, we have built OpenAQ, an innovative, open platform created by a group of scientists and open data programmers. The source code for the platform is viewable at github.com/openaq. Currently, we are aggregating, storing, and making publicly available real-time air quality data (PM2.5, PM10, SO2, NO2, and O3) via an Application Program Interface (API). We will present the OpenAQ platform, which currently has the following specific capabilities: A continuous ingest mechanism for some of the most polluted cities, generalizable to more sources An API providing data-querying, including ability to filter by location, measurement type and value and date, as well as custom sort options A generalized, chart-based visualization tool to explore data accessible via the API At this stage, we are seeking wider participation and input from multiple research communities in expanding our data retrieval sites, standardizing our protocols, receiving feedback on quality issues, and creating tools that can be built on top of this open platform.

  1. DyninstAPI Patches

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    LeGendre, M.

    2012-04-01

    We are seeking a code review of patches against DyninstAPI 8.0. DyninstAPI is an open source binary instrumentation library from the University of Wisconsin and University of Maryland. Our patches port DyninstAPI to the BlueGene/P and BlueGene/Q systems, as well as fix DyninstAPI bugs and implement minor new features in DyninstAPI.

  2. 30 CFR 250.446 - What are the BOP maintenance and inspection requirements?

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ....11, Maintenance; and Sections 17.12 and 18.12, Quality Management, described in API RP 53... 30 Mineral Resources 2 2010-07-01 2010-07-01 false What are the BOP maintenance and inspection requirements? 250.446 Section 250.446 Mineral Resources MINERALS MANAGEMENT SERVICE, DEPARTMENT OF THE INTERIOR...

  3. 77 FR 60124 - Draft Guidance for Industry on Initial Completeness Assessments for Type II Active Pharmaceutical...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-02

    ... DEPARTMENT OF HEALTH AND HUMAN SERVICES Food and Drug Administration [Docket No. FDA-2012-D-1010] Draft Guidance for Industry on Initial Completeness Assessments for Type II Active Pharmaceutical... certain drug master files, namely, Type II active pharmaceutical ingredient (API) drug master files (DMFs...

  4. 76 FR 54923 - Special Conditions: Dassault Falcon Model 900 and 900EX Airplanes; Interaction of Systems and...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-09-06

    ... Airplane Directorate, Aircraft Certification Service, 1601 Lind Avenue, SW., Renton, Washington 98057-3356... affected aircraft. In addition, the substance of these special conditions has been subject to the public... modified by API, feature a wing-load-alleviation system that precludes deployment of the air brakes at...

  5. 30 CFR 250.908 - What are the minimum structural fatigue design requirements?

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 30 Mineral Resources 2 2011-07-01 2011-07-01 false What are the minimum structural fatigue design... fatigue design requirements? (a) API RP 2A-WSD, Recommended Practice for Planning, Designing and... the design fatigue life of each joint and member be twice the intended service life of the structure...

  6. Using RxNorm and NDF-RT to classify medication data extracted from electronic health records: experiences from the Rochester Epidemiology Project.

    PubMed

    Pathak, Jyotishman; Murphy, Sean P; Willaert, Brian N; Kremers, Hilal M; Yawn, Barbara P; Rocca, Walter A; Chute, Christopher G

    2011-01-01

    RxNorm and NDF-RT published by the National Library of Medicine (NLM) and Veterans Affairs (VA), respectively, are two publicly available federal medication terminologies. In this study, we evaluate the applicability of RxNorm and National Drug File-Reference Terminology (NDF-RT) for extraction and classification of medication data retrieved using structured querying and natural language processing techniques from electronic health records at two different medical centers within the Rochester Epidemiology Project (REP). Specifically, we explore how mappings between RxNorm concept codes and NDF-RT drug classes can be leveraged for hierarchical organization and grouping of REP medication data, identify gaps and coverage issues, and analyze the recently released NLM's NDF-RT Web service API. Our study concludes that RxNorm and NDF-RT can be applied together for classification of medication extracted from multiple EHR systems, although several issues and challenges remain to be addressed. We further conclude that the Web service APIs developed by the NLM provide useful functionalities for such activities.

  7. Atlas - a data warehouse for integrative bioinformatics.

    PubMed

    Shah, Sohrab P; Huang, Yong; Xu, Tao; Yuen, Macaire M S; Ling, John; Ouellette, B F Francis

    2005-02-21

    We present a biological data warehouse called Atlas that locally stores and integrates biological sequences, molecular interactions, homology information, functional annotations of genes, and biological ontologies. The goal of the system is to provide data, as well as a software infrastructure for bioinformatics research and development. The Atlas system is based on relational data models that we developed for each of the source data types. Data stored within these relational models are managed through Structured Query Language (SQL) calls that are implemented in a set of Application Programming Interfaces (APIs). The APIs include three languages: C++, Java, and Perl. The methods in these API libraries are used to construct a set of loader applications, which parse and load the source datasets into the Atlas database, and a set of toolbox applications which facilitate data retrieval. Atlas stores and integrates local instances of GenBank, RefSeq, UniProt, Human Protein Reference Database (HPRD), Biomolecular Interaction Network Database (BIND), Database of Interacting Proteins (DIP), Molecular Interactions Database (MINT), IntAct, NCBI Taxonomy, Gene Ontology (GO), Online Mendelian Inheritance in Man (OMIM), LocusLink, Entrez Gene and HomoloGene. The retrieval APIs and toolbox applications are critical components that offer end-users flexible, easy, integrated access to this data. We present use cases that use Atlas to integrate these sources for genome annotation, inference of molecular interactions across species, and gene-disease associations. The Atlas biological data warehouse serves as data infrastructure for bioinformatics research and development. It forms the backbone of the research activities in our laboratory and facilitates the integration of disparate, heterogeneous biological sources of data enabling new scientific inferences. Atlas achieves integration of diverse data sets at two levels. First, Atlas stores data of similar types using common data models, enforcing the relationships between data types. Second, integration is achieved through a combination of APIs, ontology, and tools. The Atlas software is freely available under the GNU General Public License at: http://bioinformatics.ubc.ca/atlas/

  8. Atlas – a data warehouse for integrative bioinformatics

    PubMed Central

    Shah, Sohrab P; Huang, Yong; Xu, Tao; Yuen, Macaire MS; Ling, John; Ouellette, BF Francis

    2005-01-01

    Background We present a biological data warehouse called Atlas that locally stores and integrates biological sequences, molecular interactions, homology information, functional annotations of genes, and biological ontologies. The goal of the system is to provide data, as well as a software infrastructure for bioinformatics research and development. Description The Atlas system is based on relational data models that we developed for each of the source data types. Data stored within these relational models are managed through Structured Query Language (SQL) calls that are implemented in a set of Application Programming Interfaces (APIs). The APIs include three languages: C++, Java, and Perl. The methods in these API libraries are used to construct a set of loader applications, which parse and load the source datasets into the Atlas database, and a set of toolbox applications which facilitate data retrieval. Atlas stores and integrates local instances of GenBank, RefSeq, UniProt, Human Protein Reference Database (HPRD), Biomolecular Interaction Network Database (BIND), Database of Interacting Proteins (DIP), Molecular Interactions Database (MINT), IntAct, NCBI Taxonomy, Gene Ontology (GO), Online Mendelian Inheritance in Man (OMIM), LocusLink, Entrez Gene and HomoloGene. The retrieval APIs and toolbox applications are critical components that offer end-users flexible, easy, integrated access to this data. We present use cases that use Atlas to integrate these sources for genome annotation, inference of molecular interactions across species, and gene-disease associations. Conclusion The Atlas biological data warehouse serves as data infrastructure for bioinformatics research and development. It forms the backbone of the research activities in our laboratory and facilitates the integration of disparate, heterogeneous biological sources of data enabling new scientific inferences. Atlas achieves integration of diverse data sets at two levels. First, Atlas stores data of similar types using common data models, enforcing the relationships between data types. Second, integration is achieved through a combination of APIs, ontology, and tools. The Atlas software is freely available under the GNU General Public License at: PMID:15723693

  9. Raman spectroscopy as a PAT for pharmaceutical blending: Advantages and disadvantages.

    PubMed

    Riolo, Daniela; Piazza, Alessandro; Cottini, Ciro; Serafini, Margherita; Lutero, Emilio; Cuoghi, Erika; Gasparini, Lorena; Botturi, Debora; Marino, Iari Gabriel; Aliatis, Irene; Bersani, Danilo; Lottici, Pier Paolo

    2018-02-05

    Raman spectroscopy has been positively evaluated as a tool for the in-line and real-time monitoring of powder blending processes and it has been proved to be effective in the determination of the endpoint of the mixing, showing its potential role as process analytical technology (PAT). The aim of this study is to show advantages and disadvantages of Raman spectroscopy with respect to the most traditional HPLC analysis. The spectroscopic results, obtained directly on raw powders, sampled from a two-axis blender in real case conditions, were compared with the chromatographic data obtained on the same samples. The formulation blend used for the experiment consists of active pharmaceutical ingredient (API, concentrations 6.0% and 0.5%), lactose and magnesium stearate (as excipients). The first step of the monitoring process was selecting the appropriate wavenumber region where the Raman signal of API is maximal and interference from the spectral features of excipients is minimal. Blend profiles were created by plotting the area ratios of the Raman peak of API (A API ) at 1598cm -1 and the Raman bands of excipients (A EXC ), in the spectral range between 1560 and 1630cm -1 , as a function of mixing time: the API content can be considered homogeneous when the time-dependent dispersion of the area ratio is minimized. In order to achieve a representative sampling with Raman spectroscopy, each sample was mapped in a motorized XY stage by a defocused laser beam of a micro-Raman apparatus. Good correlation between the two techniques has been found only for the composition at 6.0% (w/w). However, standard deviation analysis, applied to both HPLC and Raman data, showed that Raman results are more substantial than HPLC ones, since Raman spectroscopy enables generating data rich blend profiles. In addition, the relative standard deviation calculated from a single map (30 points) turned out to be representative of the degree of homogeneity for that blend time. Copyright © 2017 Elsevier B.V. All rights reserved.

  10. The NASA Reanalysis Ensemble Service - Advanced Capabilities for Integrated Reanalysis Access and Intercomparison

    NASA Astrophysics Data System (ADS)

    Tamkin, G.; Schnase, J. L.; Duffy, D.; Li, J.; Strong, S.; Thompson, J. H.

    2017-12-01

    NASA's efforts to advance climate analytics-as-a-service are making new capabilities available to the research community: (1) A full-featured Reanalysis Ensemble Service (RES) comprising monthly means data from multiple reanalysis data sets, accessible through an enhanced set of extraction, analytic, arithmetic, and intercomparison operations. The operations are made accessible through NASA's climate data analytics Web services and our client-side Climate Data Services Python library, CDSlib; (2) A cloud-based, high-performance Virtual Real-Time Analytics Testbed supporting a select set of climate variables. This near real-time capability enables advanced technologies like Spark and Hadoop-based MapReduce analytics over native NetCDF files; and (3) A WPS-compliant Web service interface to our climate data analytics service that will enable greater interoperability with next-generation systems such as ESGF. The Reanalysis Ensemble Service includes the following: - New API that supports full temporal, spatial, and grid-based resolution services with sample queries - A Docker-ready RES application to deploy across platforms - Extended capabilities that enable single- and multiple reanalysis area average, vertical average, re-gridding, standard deviation, and ensemble averages - Convenient, one-stop shopping for commonly used data products from multiple reanalyses including basic sub-setting and arithmetic operations (e.g., avg, sum, max, min, var, count, anomaly) - Full support for the MERRA-2 reanalysis dataset in addition to, ECMWF ERA-Interim, NCEP CFSR, JMA JRA-55 and NOAA/ESRL 20CR… - A Jupyter notebook-based distribution mechanism designed for client use cases that combines CDSlib documentation with interactive scenarios and personalized project management - Supporting analytic services for NASA GMAO Forward Processing datasets - Basic uncertainty quantification services that combine heterogeneous ensemble products with comparative observational products (e.g., reanalysis, observational, visualization) - The ability to compute and visualize multiple reanalysis for ease of inter-comparisons - Automated tools to retrieve and prepare data collections for analytic processing

  11. A Web Service Protocol Realizing Interoperable Internet of Things Tasking Capability

    PubMed Central

    Huang, Chih-Yuan; Wu, Cheng-Hung

    2016-01-01

    The Internet of Things (IoT) is an infrastructure that interconnects uniquely-identifiable devices using the Internet. By interconnecting everyday appliances, various monitoring, and physical mashup applications can be constructed to improve human’s daily life. In general, IoT devices provide two main capabilities: sensing and tasking capabilities. While the sensing capability is similar to the World-Wide Sensor Web, this research focuses on the tasking capability. However, currently, IoT devices created by different manufacturers follow different proprietary protocols and are locked in many closed ecosystems. This heterogeneity issue impedes the interconnection between IoT devices and damages the potential of the IoT. To address this issue, this research aims at proposing an interoperable solution called tasking capability description that allows users to control different IoT devices using a uniform web service interface. This paper demonstrates the contribution of the proposed solution by interconnecting different IoT devices for different applications. In addition, the proposed solution is integrated with the OGC SensorThings API standard, which is a Web service standard defined for the IoT sensing capability. Consequently, the Extended SensorThings API can realize both IoT sensing and tasking capabilities in an integrated and interoperable manner. PMID:27589759

  12. Sharing environmental models: An Approach using GitHub repositories and Web Processing Services

    NASA Astrophysics Data System (ADS)

    Stasch, Christoph; Nuest, Daniel; Pross, Benjamin

    2016-04-01

    The GLUES (Global Assessment of Land Use Dynamics, Greenhouse Gas Emissions and Ecosystem Services) project established a spatial data infrastructure for scientific geospatial data and metadata (http://geoportal-glues.ufz.de), where different regional collaborative projects researching the impacts of climate and socio-economic changes on sustainable land management can share their underlying base scenarios and datasets. One goal of the project is to ease the sharing of computational models between institutions and to make them easily executable in Web-based infrastructures. In this work, we present such an approach for sharing computational models relying on GitHub repositories (http://github.com) and Web Processing Services. At first, model providers upload their model implementations to GitHub repositories in order to share them with others. The GitHub platform allows users to submit changes to the model code. The changes can be discussed and reviewed before merging them. However, while GitHub allows sharing and collaborating of model source code, it does not actually allow running these models, which requires efforts to transfer the implementation to a model execution framework. We thus have extended an existing implementation of the OGC Web Processing Service standard (http://www.opengeospatial.org/standards/wps), the 52°North Web Processing Service (http://52north.org/wps) platform to retrieve all model implementations from a git (http://git-scm.com) repository and add them to the collection of published geoprocesses. The current implementation is restricted to models implemented as R scripts using WPS4R annotations (Hinz et al.) and to Java algorithms using the 52°North WPS Java API. The models hence become executable through a standardized Web API by multiple clients such as desktop or browser GIS and modelling frameworks. If the model code is changed on the GitHub platform, the changes are retrieved by the service and the processes will be updated accordingly. The admin tool of the 52°North WPS was extended to support automated retrieval and deployment of computational models from GitHub repositories. Once the R code is available in the GitHub repo, the contained process can be easily deployed and executed by simply defining the GitHub repository URL in the WPS admin tool. We illustrate the usage of the approach by sharing and running a model for land use system archetypes developed by the Helmholtz Centre for Environmental Research (UFZ, see Vaclavik et al.). The original R code was extended and published in the 52°North WPS using both, public and non-public datasets (Nüst et al., see also https://github.com/52North/glues-wps). Hosting the analysis in a Git repository now allows WPS administrators, client developers, and modelers to easily work together on new versions or completely new web processes using the powerful GitHub collaboration platform. References: Hinz, M. et. al. (2013): Spatial Statistics on the Geospatial Web. In: The 16th AGILE International Conference on Geographic Information Science, Short Papers. http://www.agile-online.org/Conference_Paper/CDs/agile_2013/Short_Papers/SP_S3.1_Hinz.pdf Nüst, D. et. al.: (2015): Open and reproducible global land use classification. In: EGU General Assembly Conference Abstracts . Vol. 17. European Geophysical Union, 2015, p. 9125, http://meetingorganizer.copernicus. org/EGU2015/EGU2015- 9125.pdf Vaclavik, T., et. al. (2013): Mapping global land system archetypes. Global Environmental Change 23(6): 1637-1647. Online available: October 9, 2013, DOI: 10.1016/j.gloenvcha.2013.09.004

  13. Fusion of the C-terminal triskaidecapeptide of hirudin variant 3 to alpha1-proteinase inhibitor M358R increases the serpin-mediated rate of thrombin inhibition

    PubMed Central

    2013-01-01

    Background Alpha-1 proteinase inhibitor (API) is a plasma serpin superfamily member that inhibits neutrophil elastase; variant API M358R inhibits thrombin and activated protein C (APC). Fusing residues 1-75 of another serpin, heparin cofactor II (HCII), to API M358R (in HAPI M358R) was previously shown to accelerate thrombin inhibition over API M358R by conferring thrombin exosite 1 binding properties. We hypothesized that replacing HCII 1-75 region with the 13 C-terminal residues (triskaidecapeptide) of hirudin variant 3 (HV354-66) would further enhance the inhibitory potency of API M358R fusion proteins. We therefore expressed HV3API M358R (HV354-66 fused to API M358R) and HV3API RCL5 (HV354-66 fused to API F352A/L353V/E354V/A355I/I356A/I460L/M358R) API M358R) as N-terminally hexahistidine-tagged polypeptides in E. coli. Results HV3API M358R inhibited thrombin 3.3-fold more rapidly than API M358R; for HV3API RCL5 the rate enhancement was 1.9-fold versus API RCL5; neither protein inhibited thrombin as rapidly as HAPI M358R. While the thrombin/Activated Protein C rate constant ratio was 77-fold higher for HV3API RCL5 than for HV3API M358R, most of the increased specificity derived from the API F352A/L353V/E354V/A355I/I356A/I460L API RCL 5 mutations, since API RCL5 remained 3-fold more specific than HV3API RCL5. An HV3 54-66 peptide doubled the Thrombin Clotting Time (TCT) and halved the binding of thrombin to immobilized HCII 1-75 at lower concentrations than free HCII 1-75. HV3API RCL5 bound active site-inhibited FPR-chloromethyl ketone-thrombin more effectively than HAPI RCL5. Transferring the position of the fused HV3 triskaidecapeptide to the C-terminus of API M358R decreased the rate of thrombin inhibition relative to that mediated by HV3API M358R by 11-to 14-fold. Conclusions Fusing the C-terminal triskaidecapeptide of HV3 to API M358R-containing serpins significantly increased their effectiveness as thrombin inhibitors, but the enhancement was less than that seen in HCII 1-75–API M358R fusion proteins. HCII 1-75 was a superior fusion partner, in spite of the greater affinity of the HV3 triskaidecapeptide, manifested both in isolated and API-fused form, for thrombin exosite 1. Our results suggest that HCII 1-75 binds thrombin exosite 1 and orients the attached serpin scaffold for more efficient interaction with the active site of thrombin than the HV3 triskaidecapeptide. PMID:24215622

  14. Infrared imaging spectroscopy and chemometric tools for in situ analysis of an imiquimod pharmaceutical preparation presented as cream

    NASA Astrophysics Data System (ADS)

    Carneiro, Renato Lajarim; Poppi, Ronei Jesus

    2014-01-01

    In the present work the homogeneity of a pharmaceutical formulation presented as a cream was studied using infrared imaging spectroscopy and chemometric methodologies such as principal component analysis (PCA) and multivariate curve resolution with alternating least squares (MCR-ALS). A cream formulation, presented as an emulsion, was prepared using imiquimod as the active pharmaceutical ingredient (API) and the excipients: water, vaseline, an emulsifier and a carboxylic acid in order to dissolve the API. After exposure at 45 °C during 3 months to perform accelerated stability test, the presence of some crystals was observed, indicating homogeneity problems in the formulation. PCA exploratory analysis showed that the crystal composition was different from the composition of the emulsion, since the score maps presented crystal structures in the emulsion. MCR-ALS estimated the spectra of the crystals and the emulsion. The crystals presented amine and C-H bands, suggesting that the precipitate was a salt formed by carboxylic acid and imiquimod. These results indicate the potential of infrared imaging spectroscopy in conjunction with chemometric methodologies as an analytical tool to ensure the quality of cream formulations in the pharmaceutical industry.

  15. libNeuroML and PyLEMS: using Python to combine procedural and declarative modeling approaches in computational neuroscience.

    PubMed

    Vella, Michael; Cannon, Robert C; Crook, Sharon; Davison, Andrew P; Ganapathy, Gautham; Robinson, Hugh P C; Silver, R Angus; Gleeson, Padraig

    2014-01-01

    NeuroML is an XML-based model description language, which provides a powerful common data format for defining and exchanging models of neurons and neuronal networks. In the latest version of NeuroML, the structure and behavior of ion channel, synapse, cell, and network model descriptions are based on underlying definitions provided in LEMS, a domain-independent language for expressing hierarchical mathematical models of physical entities. While declarative approaches for describing models have led to greater exchange of model elements among software tools in computational neuroscience, a frequent criticism of XML-based languages is that they are difficult to work with directly. Here we describe two Application Programming Interfaces (APIs) written in Python (http://www.python.org), which simplify the process of developing and modifying models expressed in NeuroML and LEMS. The libNeuroML API provides a Python object model with a direct mapping to all NeuroML concepts defined by the NeuroML Schema, which facilitates reading and writing the XML equivalents. In addition, it offers a memory-efficient, array-based internal representation, which is useful for handling large-scale connectomics data. The libNeuroML API also includes support for performing common operations that are required when working with NeuroML documents. Access to the LEMS data model is provided by the PyLEMS API, which provides a Python implementation of the LEMS language, including the ability to simulate most models expressed in LEMS. Together, libNeuroML and PyLEMS provide a comprehensive solution for interacting with NeuroML models in a Python environment.

  16. libNeuroML and PyLEMS: using Python to combine procedural and declarative modeling approaches in computational neuroscience

    PubMed Central

    Vella, Michael; Cannon, Robert C.; Crook, Sharon; Davison, Andrew P.; Ganapathy, Gautham; Robinson, Hugh P. C.; Silver, R. Angus; Gleeson, Padraig

    2014-01-01

    NeuroML is an XML-based model description language, which provides a powerful common data format for defining and exchanging models of neurons and neuronal networks. In the latest version of NeuroML, the structure and behavior of ion channel, synapse, cell, and network model descriptions are based on underlying definitions provided in LEMS, a domain-independent language for expressing hierarchical mathematical models of physical entities. While declarative approaches for describing models have led to greater exchange of model elements among software tools in computational neuroscience, a frequent criticism of XML-based languages is that they are difficult to work with directly. Here we describe two Application Programming Interfaces (APIs) written in Python (http://www.python.org), which simplify the process of developing and modifying models expressed in NeuroML and LEMS. The libNeuroML API provides a Python object model with a direct mapping to all NeuroML concepts defined by the NeuroML Schema, which facilitates reading and writing the XML equivalents. In addition, it offers a memory-efficient, array-based internal representation, which is useful for handling large-scale connectomics data. The libNeuroML API also includes support for performing common operations that are required when working with NeuroML documents. Access to the LEMS data model is provided by the PyLEMS API, which provides a Python implementation of the LEMS language, including the ability to simulate most models expressed in LEMS. Together, libNeuroML and PyLEMS provide a comprehensive solution for interacting with NeuroML models in a Python environment. PMID:24795618

  17. Tactical Web Services: Using XML and Java Web Services to Conduct Real-Time Net-Centric Sonar Visualization

    DTIC Science & Technology

    2004-09-01

    Rosetti USN U.S. Navy Chesterton, IN 6. Erik Chaum NUWC Newport, RI 7. David Bellino NPRI Newport, RI 8. Dick Nadolink NUWC Newport, RI...found at (http://www.parallelgraphics.com/products/cortona). G. JFREECHART JFreeChart is an open source Java API created by David Gilbert and...www.xj3d.org/. Accessed 3 September 2004. Hunter, David , Kurt Cagle, and Chris Dix, eds. Beginning XML, Second Edition. Indianapolis, IN

  18. Chemozart: a web-based 3D molecular structure editor and visualizer platform.

    PubMed

    Mohebifar, Mohamad; Sajadi, Fatemehsadat

    2015-01-01

    Chemozart is a 3D Molecule editor and visualizer built on top of native web components. It offers an easy to access service, user-friendly graphical interface and modular design. It is a client centric web application which communicates with the server via a representational state transfer style web service. Both client-side and server-side application are written in JavaScript. A combination of JavaScript and HTML is used to draw three-dimensional structures of molecules. With the help of WebGL, three-dimensional visualization tool is provided. Using CSS3 and HTML5, a user-friendly interface is composed. More than 30 packages are used to compose this application which adds enough flexibility to it to be extended. Molecule structures can be drawn on all types of platforms and is compatible with mobile devices. No installation is required in order to use this application and it can be accessed through the internet. This application can be extended on both server-side and client-side by implementing modules in JavaScript. Molecular compounds are drawn on the HTML5 Canvas element using WebGL context. Chemozart is a chemical platform which is powerful, flexible, and easy to access. It provides an online web-based tool used for chemical visualization along with result oriented optimization for cloud based API (application programming interface). JavaScript libraries which allow creation of web pages containing interactive three-dimensional molecular structures has also been made available. The application has been released under Apache 2 License and is available from the project website https://chemozart.com.

  19. N-Terminal Ile-Orn- and Trp-Orn-Motif Repeats Enhance Membrane Interaction and Increase the Antimicrobial Activity of Apidaecins against Pseudomonas aeruginosa

    PubMed Central

    Bluhm, Martina E. C.; Schneider, Viktoria A. F.; Schäfer, Ingo; Piantavigna, Stefania; Goldbach, Tina; Knappe, Daniel; Seibel, Peter; Martin, Lisandra L.; Veldhuizen, Edwin J. A.; Hoffmann, Ralf

    2016-01-01

    The Gram-negative bacterium Pseudomonas aeruginosa is a life-threatening nosocomial pathogen due to its generally low susceptibility toward antibiotics. Furthermore, many strains have acquired resistance mechanisms requiring new antimicrobials with novel mechanisms to enhance treatment options. Proline-rich antimicrobial peptides, such as the apidaecin analog Api137, are highly efficient against various Enterobacteriaceae infections in mice, but less active against P. aeruginosa in vitro. Here, we extended our recent work by optimizing lead peptides Api755 (gu-OIORPVYOPRPRPPHPRL-OH; gu = N,N,N′,N′-tetramethylguanidino, O = L-ornithine) and Api760 (gu-OWORPVYOPRPRPPHPRL-OH) by incorporation of Ile-Orn- and Trp-Orn-motifs, respectively. Api795 (gu-O(IO)2RPVYOPRPRPPHPRL-OH) and Api794 (gu-O(WO)3RPVYOPRPRPPHPRL-OH) were highly active against P. aeruginosa with minimal inhibitory concentrations of 8–16 and 8–32 μg/mL against Escherichia coli and Klebsiella pneumoniae. Assessed using a quartz crystal microbalance, these peptides inserted into a membrane layer and the surface activity increased gradually from Api137, over Api795, to Api794. This mode of action was confirmed by transmission electron microscopy indicating some membrane damage only at the high peptide concentrations. Api794 and Api795 were highly stable against serum proteases (half-life times >5 h) and non-hemolytic to human erythrocytes at peptide concentrations of 0.6 g/L. At this concentration, Api795 reduced the cell viability of HeLa cells only slightly, whereas the IC50 of Api794 was 0.23 ± 0.09 g/L. Confocal fluorescence microscopy revealed no colocalization of 5(6)-carboxyfluorescein-labeled Api794 or Api795 with the mitochondria, excluding interactions with the mitochondrial membrane. Interestingly, Api795 was localized in endosomes, whereas Api794 was present in endosomes and the cytosol. This was verified using flow cytometry showing a 50% higher uptake of Api794 in HeLa cells compared with Api795. The uptake was reduced for both peptides by 50 and 80%, respectively, after inhibiting endocytotic uptake with dynasore. In summary, Api794 and Api795 were highly active against P. aeruginosa in vitro. Both peptides passed across the bacterial membrane efficiently, most likely then disturbing the ribosome assembly, and resulting in further intracellular damage. Api795 with its IOIO-motif, which was particularly active and only slightly toxic in vitro, appears to represent a promising third generation lead compound for the development of novel antibiotics against P. aeruginosa. PMID:27243004

  20. Towards structured sharing of raw and derived neuroimaging data across existing resources

    PubMed Central

    Keator, D.B.; Helmer, K.; Steffener, J.; Turner, J.A.; Van Erp, T.G.M.; Gadde, S.; Ashish, N.; Burns, G.A.; Nichols, B.N.

    2013-01-01

    Data sharing efforts increasingly contribute to the acceleration of scientific discovery. Neuroimaging data is accumulating in distributed domain-specific databases and there is currently no integrated access mechanism nor an accepted format for the critically important meta-data that is necessary for making use of the combined, available neuroimaging data. In this manuscript, we present work from the Derived Data Working Group, an open-access group sponsored by the Biomedical Informatics Research Network (BIRN) and the International Neuroimaging Coordinating Facility (INCF) focused on practical tools for distributed access to neuroimaging data. The working group develops models and tools facilitating the structured interchange of neuroimaging meta-data and is making progress towards a unified set of tools for such data and meta-data exchange. We report on the key components required for integrated access to raw and derived neuroimaging data as well as associated meta-data and provenance across neuroimaging resources. The components include (1) a structured terminology that provides semantic context to data, (2) a formal data model for neuroimaging with robust tracking of data provenance, (3) a web service-based application programming interface (API) that provides a consistent mechanism to access and query the data model, and (4) a provenance library that can be used for the extraction of provenance data by image analysts and imaging software developers. We believe that the framework and set of tools outlined in this manuscript have great potential for solving many of the issues the neuroimaging community faces when sharing raw and derived neuroimaging data across the various existing database systems for the purpose of accelerating scientific discovery. PMID:23727024

  1. What does it take to build a medium scale scientific cloud to process significant amounts of Earth observation data?

    NASA Astrophysics Data System (ADS)

    Hollstein, André; Diedrich, Hannes; Spengler, Daniel

    2017-04-01

    The installment of the operational fleet of Sentinels by Copernicus offers an unprecedented influx of freely available Earth Observation data with Sentinel-2 being a great example. It offers a broad range of land applications due to its high spatial sampling from 10 m to 20 m and its multi-spectral imaging capabilities with 13 spectral bands. The open access policy allows unrestricted use by everybody and provides data downloads for on the respective sites. For a small area of interest and shorter time series, data processing, and exploitation can easily be done manually. However, for multi-temporal analysis of larger areas, the data size can quickly increase such that it is not manageable in practice on a personal computer which leads to an increasing interest in central data exploitation platforms. Prominent examples are GoogleEarth Engine, NASA Earth Exchange (NEX) or current developments such as CODE-DE in Germany. Open standards are still evolving, and the choice of a platform may create lock-in scenarios and a situation where scientists are not anymore in full control of all aspects of their analysis. Securing intellectual properties of researchers can become a major issue in the future. Partnering with a startup company that is dedicated to providing tools for farm management and precision farming, GFZ builds a small-scale science cloud named GTS2 for processing and distribution of Sentinel-2 data. The service includes a sophisticated atmospheric correction algorithm, spatial co-registration of time series data, as well as a web API for data distribution. This approach is different from the drag to centralized research using infrastructures controlled by others. By keeping the full licensing rights, it allows developing new business models independent from the initially chosen processing provider. Currently, data is held for the greater German area but is extendable to larger areas on short notice due to a scalable distributed network file system. For a given area of interest, band and time range selection, the API returns only the data that was requested in a fast manner and thereby saves storage space on the user's machine. A jupyterhub instance is a main tool for data exploitation by our users. Nearly all used software is open source, is based on open standards, and allows to transfer software to other infrastructures. In the talk, we give an overview of the current status of the project and the service, but also want to share our experience with its development.

  2. The Error Reporting in the ATLAS TDAQ System

    NASA Astrophysics Data System (ADS)

    Kolos, Serguei; Kazarov, Andrei; Papaevgeniou, Lykourgos

    2015-05-01

    The ATLAS Error Reporting provides a service that allows experts and shift crew to track and address errors relating to the data taking components and applications. This service, called the Error Reporting Service (ERS), gives to software applications the opportunity to collect and send comprehensive data about run-time errors, to a place where it can be intercepted in real-time by any other system component. Other ATLAS online control and monitoring tools use the ERS as one of their main inputs to address system problems in a timely manner and to improve the quality of acquired data. The actual destination of the error messages depends solely on the run-time environment, in which the online applications are operating. When an application sends information to ERS, depending on the configuration, it may end up in a local file, a database, distributed middleware which can transport it to an expert system or display it to users. Thanks to the open framework design of ERS, new information destinations can be added at any moment without touching the reporting and receiving applications. The ERS Application Program Interface (API) is provided in three programming languages used in the ATLAS online environment: C++, Java and Python. All APIs use exceptions for error reporting but each of them exploits advanced features of a given language to simplify the end-user program writing. For example, as C++ lacks language support for exceptions, a number of macros have been designed to generate hierarchies of C++ exception classes at compile time. Using this approach a software developer can write a single line of code to generate a boilerplate code for a fully qualified C++ exception class declaration with arbitrary number of parameters and multiple constructors, which encapsulates all relevant static information about the given type of issues. When a corresponding error occurs at run time, the program just need to create an instance of that class passing relevant values to one of the available class constructors and send this instance to ERS. This paper presents the original design solutions exploited for the ERS implementation and describes how it was used during the first ATLAS run period. The cross-system error reporting standardization introduced by ERS was one of the key points for the successful implementation of automated mechanisms for online error recovery.

  3. The Earth System Documentation (ES-DOC) Software Process

    NASA Astrophysics Data System (ADS)

    Greenslade, M. A.; Murphy, S.; Treshansky, A.; DeLuca, C.; Guilyardi, E.; Denvil, S.

    2013-12-01

    Earth System Documentation (ES-DOC) is an international project supplying high-quality tools & services in support of earth system documentation creation, analysis and dissemination. It is nurturing a sustainable standards based documentation eco-system that aims to become an integral part of the next generation of exa-scale dataset archives. ES-DOC leverages open source software, and applies a software development methodology that places end-user narratives at the heart of all it does. ES-DOC has initially focused upon nurturing the Earth System Model (ESM) documentation eco-system and currently supporting the following projects: * Coupled Model Inter-comparison Project Phase 5 (CMIP5); * Dynamical Core Model Inter-comparison Project (DCMIP); * National Climate Predictions and Projections Platforms Quantitative Evaluation of Downscaling Workshop. This talk will demonstrate that ES-DOC implements a relatively mature software development process. Taking a pragmatic Agile process as inspiration, ES-DOC: * Iteratively develops and releases working software; * Captures user requirements via a narrative based approach; * Uses online collaboration tools (e.g. Earth System CoG) to manage progress; * Prototypes applications to validate their feasibility; * Leverages meta-programming techniques where appropriate; * Automates testing whenever sensibly feasible; * Streamlines complex deployments to a single command; * Extensively leverages GitHub and Pivotal Tracker; * Enforces strict separation of the UI from underlying API's; * Conducts code reviews.

  4. Impact of Nosema ceranae and Nosema apis on individual worker bees of the two host species (Apis cerana and Apis mellifera) and regulation of host immune response.

    PubMed

    Sinpoo, Chainarong; Paxton, Robert J; Disayathanoowat, Terd; Krongdang, Sasiprapa; Chantawannakul, Panuwan

    Nosema apis and Nosema ceranae are obligate intracellular microsporidian parasites infecting midgut epithelial cells of host adult honey bees, originally Apis mellifera and Apis cerana respectively. Each microsporidia cross-infects the other host and both microsporidia nowadays have a worldwide distribution. In this study, cross-infection experiments using both N. apis and N. ceranae in both A. mellifera and A. cerana were carried out to compare pathogen proliferation and impact on hosts, including host immune response. Infection by N. ceranae led to higher spore loads than by N. apis in both host species, and there was greater proliferation of microsporidia in A. mellifera compared to A. cerana. Both N. apis and N. ceranae were pathogenic in both host Apis species. N. ceranae induced subtly, though not significantly, higher mortality than N. apis in both host species, yet survival of A. cerana was no different to that of A. mellifera in response to N. apis or N. ceranae. Infections of both host species with N. apis and N. ceranae caused significant up-regulation of AMP genes and cellular mediated immune genes but did not greatly alter apoptosis-related gene expression. In this study, A. cerana enlisted a higher immune response and displayed lower loads of N. apis and N. ceranae spores than A. mellifera, suggesting it may be better able to defend itself against microsporidia infection. We caution against over-interpretation of our results, though, because differences between host and parasite species in survival were insignificant and because size differences between microsporidia species and between host Apis species may alternatively explain the differential proliferation of N. ceranae in A. mellifera. Copyright © 2017 Elsevier Ltd. All rights reserved.

  5. Comparison of plasma amino acid profile-based index and CA125 in the diagnosis of epithelial ovarian cancers and borderline malignant tumors.

    PubMed

    Miyagi, Etsuko; Maruyama, Yasuyo; Mogami, Tae; Numazaki, Reiko; Ikeda, Atsuko; Yamamoto, Hiroshi; Hirahara, Fumiki

    2017-02-01

    We previously developed a new plasma amino acid profile-based index (API) to detect ovarian, cervical, and endometrial cancers. Here, we compared API to serum cancer antigen 125 (CA125) for distinguishing epithelial ovarian malignant tumors from benign growths. API and CA125 were measured preoperatively in patients with ovarian tumors, which were later classified into 59 epithelial ovarian cancers, 21 epithelial borderline malignant tumors, and 97 benign tumors including 40 endometriotic cysts. The diagnostic accuracy and cutoff points of API were evaluated using receiver operating characteristic (ROC) curves. The area under the ROC curves showed the equivalent performance of API and CA125 to discriminate between malignant/borderline malignant and benign tumors (both 0.77), and API was superior to CA125 for discrimination between malignant/borderline malignant lesions and endometriotic cysts (API, 0.75 vs. CA125, 0.59; p < 0.05). At the API cutoff level of 6.0, API and CA125 had equal positive rates of detecting cancers and borderline malignancies (API, 0.71 vs. CA125, 0.74; p = 0.84) or cancers alone (API, 0.73 vs. CA125, 0.85; p = 0.12). However, API had a significantly lower detection rate of benign endometriotic cysts (0.35; 95 % CI, 0.21-0.52) compared with that of CA125 (0.65; 95 % CI, 0.48-0.79) (p < 0.05). API is an effective new tumor marker to detect ovarian cancers and borderline malignancies with a low false-positive rate for endometriosis. A large-scale prospective clinical study using the cutoff value of API determined in this study is warranted to validate API for practical clinical use.

  6. Sharing Lessons-Learned on Effective Open Data, Open-Source Practices from OpenAQ, a Global Open Air Quality Community.

    NASA Astrophysics Data System (ADS)

    Hasenkopf, C. A.

    2017-12-01

    Increasingly, open data, open-source projects are unearthing rich datasets and tools, previously impossible for more traditional avenues to generate. These projects are possible, in part, because of the emergence of online collaborative and code-sharing tools, decreasing costs of cloud-based services to fetch, store, and serve data, and increasing interest of individuals to contribute their time and skills to 'open projects.' While such projects have generated palpable enthusiasm from many sectors, many of these projects face uncharted paths for sustainability, visibility, and acceptance. Our project, OpenAQ, is an example of an open-source, open data community that is currently forging its own uncharted path. OpenAQ is an open air quality data platform that aggregates and universally formats government and research-grade air quality data from 50 countries across the world. To date, we make available more than 76 million air quality (PM2.5, PM10, SO2, NO2, O3, CO and black carbon) data points through an open Application Programming Interface (API) and a user-customizable download interface at https://openaq.org. The goal of the platform is to enable an ecosystem of users to advance air pollution efforts from science to policy to the private sector. The platform is also an open-source project (https://github.com/openaq) and has only been made possible through the coding and data contributions of individuals around the world. In our first two years of existence, we have seen requests for data to our API skyrocket to more than 6 million datapoints per month, and use-cases as varied as ingesting data aggregated from our system into real-time models of wildfires to building open-source statistical packages (e.g. ropenaq and py-openaq) on top of the platform to creating public-friendly apps and chatbots. We will share a whirl-wind trip through our evolution and the many lessons learned so far related to platform structure, community engagement, organizational model type and sustainability.

  7. The HydroShare Collaborative Repository for the Hydrology Community

    NASA Astrophysics Data System (ADS)

    Tarboton, D. G.; Idaszak, R.; Horsburgh, J. S.; Ames, D. P.; Goodall, J. L.; Couch, A.; Hooper, R. P.; Dash, P. K.; Stealey, M.; Yi, H.; Bandaragoda, C.; Castronova, A. M.

    2017-12-01

    HydroShare is an online, collaboration system for sharing of hydrologic data, analytical tools, and models. It supports the sharing of, and collaboration around, "resources" which are defined by standardized content types for data formats and models commonly used in hydrology. With HydroShare you can: Share your data and models with colleagues; Manage who has access to the content that you share; Share, access, visualize and manipulate a broad set of hydrologic data types and models; Use the web services application programming interface (API) to program automated and client access; Publish data and models and obtain a citable digital object identifier (DOI); Aggregate your resources into collections; Discover and access data and models published by others; Use web apps to visualize, analyze and run models on data in HydroShare. This presentation will describe the functionality and architecture of HydroShare highlighting our approach to making this system easy to use and serving the needs of the hydrology community represented by the Consortium of Universities for the Advancement of Hydrologic Sciences, Inc. (CUAHSI). Metadata for uploaded files is harvested automatically or captured using easy to use web user interfaces. Users are encouraged to add or create resources in HydroShare early in the data life cycle. To encourage this we allow users to share and collaborate on HydroShare resources privately among individual users or groups, entering metadata while doing the work. HydroShare also provides enhanced functionality for users through web apps that provide tools and computational capability for actions on resources. HydroShare's architecture broadly is comprised of: (1) resource storage, (2) resource exploration website, and (3) web apps for actions on resources. System components are loosely coupled and interact through APIs, which enhances robustness, as components can be upgraded and advanced relatively independently. The full power of this paradigm is the extensibility it supports. Web apps are hosted on separate servers, which may be 3rd party servers. They are registered in HydroShare using a web app resource that configures the connectivity for them to be discovered and launched directly from resource types they are associated with.

  8. Future Standardization of Space Telecommunications Radio System with Core Flight System

    NASA Technical Reports Server (NTRS)

    Hickey, Joseph P.; Briones, Janette C.; Roche, Rigoberto; Handler, Louis M.; Hall, Steven

    2016-01-01

    NASA Glenn Research Center (GRC) is integrating the NASA Space Telecommunications Radio System (STRS) Standard with the Core Flight System (cFS). The STRS standard provides a common, consistent framework to develop, qualify, operate and maintain complex, reconfigurable and reprogrammable radio systems. The cFS is a flexible, open architecture that features a plug-and-play software executive called the Core Flight Executive (cFE), a reusable library of software components for flight and space missions and an integrated tool suite. Together, STRS and cFS create a development environment that allows for STRS compliant applications to reference the STRS APIs through the cFS infrastructure. These APis are used to standardize the communication protocols on NASAs space SDRs. The cFE-STRS Operating Environment (OE) is a portable cFS library, which adds the ability to run STRS applications on existing cFS platforms. The purpose of this paper is to discuss the cFE-STRS OE prototype, preliminary experimental results performed using the Advanced Space Radio Platform (ASRP), the GRC Sband Ground Station and the SCaN (Space Communication and Navigation) Testbed currently flying onboard the International Space Station. Additionally, this paper presents a demonstration of the Consultative Committee for Space Data Systems (CCSDS) Spacecraft Onboard Interface Services (SOIS) using electronic data sheets inside cFE. This configuration allows for the data sheets to specify binary formats for data exchange between STRS applications. The integration of STRS with cFS leverages mission-proven platform functions and mitigates barriers to integration with future missions. This reduces flight software development time and the costs of software-defined radio (SDR) platforms. Furthermore, the combined benefits of STRS standardization with the flexibility of cFS provide an effective, reliable and modular framework to minimize software development efforts for spaceflight missions.

  9. Active pharmaceutical ingredients for antiretroviral treatment in low- and middle-income countries: a survey.

    PubMed

    Fortunak, Joseph M; de Souza, Rodrigo O M A; Kulkarni, Amol A; King, Christopher L; Ellison, Tiffany; Miranda, Leandro S M

    2014-01-01

    Active pharmaceutical ingredients (APIs) are the molecular entities that exert the therapeutic effects of medicines. This article provides an overview of the major APIs that are entered into antiretroviral therapy (ART), outlines how APIs are manufactured, and examines the regulatory and cost frameworks of manufacturing ART APIs used in low- and middle-income countries (LMICs). Almost all APIs for ART are prepared by chemical synthesis. Roughly 15 APIs account for essentially all of the ARTs used in LMICs. Nearly all of the ART APIs purchased through the Global Fund for AIDS, TB and Malaria (GFATM) or the United States President's Emergency Plan for AIDS Relief (PEPFAR) are produced by generic companies. API costs are very important because they are the largest contribution to the overall cost of ART. Efficient API production requires substantial investment in chemical manufacturing technologies and the ready availability of raw materials and energy at competitive prices. Generic API production is practiced in only a limited number of countries; the API market for ART is dominated by Indian companies. The quality of these APIs is ensured by manufacturing under good manufacturing practice (GMP), including process validation, testing against previously established specifications and the demonstration of clinical bioequivalence. The investment and personnel costs of a quality management system for GMP contribute significantly to the cost of API production. Chinese companies are the major suppliers for many advanced intermediates in API production. Improved chemistry of manufacturing, economies of scale and optimization of procurement have enabled drastic cost reductions for many ART APIs. The available capacity for global production of quality-assured APIs is likely adequate to meet forecasted demand for 2015. The increased use of ART for paediatric treatment, for second-line and salvage therapy, and the introduction of new APIs and combinations are important factors for the future of treatment in LMICs. The introduction of new fixed-dose combinations for ART and use of new drug delivery technologies could plausibly provide robust, durable ART for all patients in need, at an overall cost that is only moderately higher than what is presently being spent.

  10. Active pharmaceutical ingredients for antiretroviral treatment in low- and middle-income countries: a survey

    PubMed Central

    Fortunak, Joseph M; de Souza, Rodrigo OMA; Kulkarni, Amol A; King, Christopher L; Ellison, Tiffany; Miranda, Leandro SM

    2015-01-01

    Active pharmaceutical ingredients (APIs) are the molecular entities that exert the therapeutic effects of medicines. This article provides an overview of the major APIs that are entered into antiretroviral therapy (ART), outlines how APIs are manufactured, and examines the regulatory and cost frameworks of manufacturing ART APIs used in low- and middle-income countries (LMICs). Almost all APIs for ART are prepared by chemical synthesis. Roughly 15 APIs account for essentially all of the ARTs used in LMICs. Nearly all of the ART APIs purchased through the Global Fund for AIDS, TB and Malaria (GFATM) or the United States President’s Emergency Plan for AIDS Relief (PEPFAR) are produced by generic companies. API costs are very important because they are the largest contribution to the overall cost of ART. Efficient API production requires substantial investment in chemical manufacturing technologies and the ready availability of raw materials and energy at competitive prices. Generic API production is practiced in only a limited number of countries; the API market for ART is dominated by Indian companies. The quality of these APIs is ensured by manufacturing under good manufacturing practice (GMP), including process validation, testing against previously established specifications and the demonstration of clinical bioequivalence. The investment and personnel costs of a quality management system for GMP contribute significantly to the cost of API production. Chinese companies are the major suppliers for many advanced intermediates in API production. Improved chemistry of manufacturing, economies of scale and optimization of procurement have enabled drastic cost reductions for many ART APIs. The available capacity for global production of quality-assured APIs is likely adequate to meet forecasted demand for 2015. The increased use of ART for paediatric treatment, for second-line and salvage therapy, and the introduction of new APIs and combinations are important factors for the future of treatment in LMICs. The introduction of new fixed-dose combinations for ART and use of new drug delivery technologies could plausibly provide robust, durable ART for all patients in need, at an overall cost that is only moderately higher than what is presently being spent. PMID:25310430

  11. ATLAS@AWS

    NASA Astrophysics Data System (ADS)

    Gehrcke, Jan-Philip; Kluth, Stefan; Stonjek, Stefan

    2010-04-01

    We show how the ATLAS offline software is ported on the Amazon Elastic Compute Cloud (EC2). We prepare an Amazon Machine Image (AMI) on the basis of the standard ATLAS platform Scientific Linux 4 (SL4). Then an instance of the SLC4 AMI is started on EC2 and we install and validate a recent release of the ATLAS offline software distribution kit. The installed software is archived as an image on the Amazon Simple Storage Service (S3) and can be quickly retrieved and connected to new SL4 AMI instances using the Amazon Elastic Block Store (EBS). ATLAS jobs can then configure against the release kit using the ATLAS configuration management tool (cmt) in the standard way. The output of jobs is exported to S3 before the SL4 AMI is terminated. Job status information is transferred to the Amazon SimpleDB service. The whole process of launching instances of our AMI, starting, monitoring and stopping jobs and retrieving job output from S3 is controlled from a client machine using python scripts implementing the Amazon EC2/S3 API via the boto library working together with small scripts embedded in the SL4 AMI. We report our experience with setting up and operating the system using standard ATLAS job transforms.

  12. Providing Access to a Diverse Set of Global Reanalysis Dataset Collections

    NASA Astrophysics Data System (ADS)

    Schuster, D.; Worley, S. J.

    2015-12-01

    The National Center for Atmospheric Research (NCAR) Research Data Archive (RDA, http://rda.ucar.edu) provides open access to a variety of global reanalysis dataset collections to support atmospheric and related sciences research worldwide. These include products from the European Centre for Medium-Range Weather Forecasts (ECMWF), Japan Meteorological Agency (JMA), National Centers for Environmental Prediction (NCEP), National Oceanic and Atmospheric Administration (NOAA), and NCAR.All RDA hosted reanalysis collections are freely accessible to registered users through a variety of methods. Standard access methods include traditional browser and scripted HTTP file download. Enhanced downloads are available through the Globus GridFTP "fire and forget" data transfer service, which provides an efficient, reliable, and preferred alternative to traditional HTTP-based methods. For those that favor interoperable access using compatible tools, the Unidata THREDDS Data server provides remote access to complete reanalysis collections by virtual dataset aggregation "files". Finally, users can request data subsets and format conversions to be prepared for them through web interface form requests or web service API batch requests. This approach uses NCAR HPC and central file systems to effectively prepare products from the high-resolution and very large reanalyses archives. The presentation will include a detailed inventory of all RDA reanalysis dataset collection holdings, and highlight access capabilities to these collections through use case examples.

  13. Quality assessment of pharmaceutical tablet samples using Fourier transform near infrared spectroscopy and multivariate analysis

    NASA Astrophysics Data System (ADS)

    Kandpal, Lalit Mohan; Tewari, Jagdish; Gopinathan, Nishanth; Stolee, Jessica; Strong, Rick; Boulas, Pierre; Cho, Byoung-Kwan

    2017-09-01

    Determination of the content uniformity, assessed by the amount of an active pharmaceutical ingredient (API), and hardness of pharmaceutical materials is important for achieving a high-quality formulation and to ensure the intended therapeutic effects of the end-product. In this work, Fourier transform near infrared (FT-NIR) spectroscopy was used to determine the content uniformity and hardness of a pharmaceutical mini-tablet and standard tablet samples. Tablet samples were scanned using an FT-NIR instrument and tablet spectra were collected at wavelengths of 1000-2500 nm. Furthermore, multivariate analysis was applied to extract the relationship between the FT-NIR spectra and the measured parameters. The results of FT-NIR spectroscopy for API and hardness prediction were as precise as the reference high-performance liquid chromatography and mechanical hardness tests. For the prediction of mini-tablet API content, the highest coefficient of determination for the prediction (R2p) was found to be 0.99 with a standard error of prediction (SEP) of 0.72 mg. Moreover, the standard tablet hardness measurement had a R2p value of 0.91 with an SEP of 0.25 kg. These results suggest that FT-NIR spectroscopy is an alternative and accurate nondestructive measurement tool for the detection of the chemical and physical properties of pharmaceutical samples.

  14. Genetic Engineering of Bee Gut Microbiome Bacteria with a Toolkit for Modular Assembly of Broad-Host-Range Plasmids.

    PubMed

    Leonard, Sean P; Perutka, Jiri; Powell, J Elijah; Geng, Peng; Richhart, Darby D; Byrom, Michelle; Kar, Shaunak; Davies, Bryan W; Ellington, Andrew D; Moran, Nancy A; Barrick, Jeffrey E

    2018-05-18

    Engineering the bacteria present in animal microbiomes promises to lead to breakthroughs in medicine and agriculture, but progress is hampered by a dearth of tools for genetically modifying the diverse species that comprise these communities. Here we present a toolkit of genetic parts for the modular construction of broad-host-range plasmids built around the RSF1010 replicon. Golden Gate assembly of parts in this toolkit can be used to rapidly test various antibiotic resistance markers, promoters, fluorescent reporters, and other coding sequences in newly isolated bacteria. We demonstrate the utility of this toolkit in multiple species of Proteobacteria that are native to the gut microbiomes of honey bees ( Apis mellifera) and bumble bees (B ombus sp.). Expressing fluorescent proteins in Snodgrassella alvi, Gilliamella apicola, Bartonella apis, and Serratia strains enables us to visualize how these bacteria colonize the bee gut. We also demonstrate CRISPRi repression in B. apis and use Cas9-facilitated knockout of an S. alvi adhesion gene to show that it is important for colonization of the gut. Beyond characterizing how the gut microbiome influences the health of these prominent pollinators, this bee microbiome toolkit (BTK) will be useful for engineering bacteria found in other natural microbial communities.

  15. A Scientific Data Provenance API for Distributed Applications

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Raju, Bibi; Elsethagen, Todd O.; Stephan, Eric G.

    Data provenance has been an active area of research as a means to standardize how the origin of data, process event history, and what or who was responsible for influencing results is explained. There are two approaches to capture provenance information. The first approach is to collect observed evidence produced by an executing application using log files, event listeners, and temporary files that are used by the application or application developer. The provenance translated from these observations is an interpretation of the provided evidence. The second approach is called disclosed because the application provides a firsthand account of the provenancemore » based on the anticipated questions on data flow, process flow, and responsible agents. Most observed provenance collection systems collect lot of provenance information during an application run or workflow execution. The common trend in capturing provenance is to collect all possible information, then attempt to find relevant information, which is not efficient. Existing disclosed provenance system APIs do not work well in distributed environment and have trouble finding where to fit the individual pieces of provenance information. This work focuses on determining more reliable solutions for provenance capture. As part of the Integrated End-to-end Performance Prediction and Diagnosis for Extreme Scientific Workflows (IPPD) project, an API was developed, called Producer API (PAPI), which can disclose application targeted provenance, designed to work in distributed environments by means of unique object identification methods. The provenance disclosure approach used adds additional metadata to the provenance information to uniquely identify the pieces and connect them together. PAPI uses a common provenance model to support this provenance integration across disclosure sources. The API also provides the flexibility to let the user decide what to do with the collected provenance. The collected provenance can be sent to a triple store using REST services or it can be logged to a file.« less

  16. The Best of all Possible Worlds: Applying the Model Driven Architecture Approach to a JC3IEDM OWL Ontology Modeled in UML

    DTIC Science & Technology

    2014-04-25

    EA’s Java application programming interface (API), the team built a tool called OWL2EA that can ingest an OWL file and generate the corresponding UML...ObjectItemStructure specification shown in Figure 10. Running this script in the relational database server MySQL creates the physical schema that

  17. MemAxes Visualization Software

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hardware advancements such as Intel's PEBS and AMD's IBS, as well as software developments such as the perf_event API in Linux have made available the acquisition of memory access samples with performance information. MemAxes is a visualization and analysis tool for memory access sample data. By mapping the samples to their associated code, variables, node topology, and application dataset, MemAxes provides intuitive views of the data.

  18. Bitterness prediction in-silico: A step towards better drugs.

    PubMed

    Bahia, Malkeet Singh; Nissim, Ido; Niv, Masha Y

    2018-02-05

    Bitter taste is innately aversive and thought to protect against consuming poisons. Bitter taste receptors (Tas2Rs) are G-protein coupled receptors, expressed both orally and extra-orally and proposed as novel targets for several indications, including asthma. Many clinical drugs elicit bitter taste, suggesting the possibility of drugs re-purposing. On the other hand, the bitter taste of medicine presents a major compliance problem for pediatric drugs. Thus, efficient tools for predicting, measuring and masking bitterness of active pharmaceutical ingredients (APIs) are required by the pharmaceutical industry. Here we highlight the BitterDB database of bitter compounds and survey the main computational approaches to prediction of bitter taste based on compound's chemical structure. Current in silico bitterness prediction methods provide encouraging results, can be constantly improved using growing experimental data, and present a reliable and efficient addition to the APIs development toolbox. Copyright © 2017 Elsevier B.V. All rights reserved.

  19. Updates to the NASA Space Telecommunications Radio System (STRS) Architecture

    NASA Technical Reports Server (NTRS)

    Kacpura, Thomas J.; Handler, Louis M.; Briones, Janette; Hall, Charles S.

    2008-01-01

    This paper describes an update of the Space Telecommunications Radio System (STRS) open architecture for NASA space based radios. The STRS architecture has been defined as a framework for the design, development, operation and upgrade of space based software defined radios, where processing resources are constrained. The architecture has been updated based upon reviews by NASA missions, radio providers, and component vendors. The STRS Standard prescribes the architectural relationship between the software elements used in software execution and defines the Application Programmer Interface (API) between the operating environment and the waveform application. Modeling tools have been adopted to present the architecture. The paper will present a description of the updated API, configuration files, and constraints. Minimum compliance is discussed for early implementations. The paper then closes with a summary of the changes made and discussion of the relevant alignment with the Object Management Group (OMG) SWRadio specification, and enhancements to the specialized signal processing abstraction.

  20. Radio Frequency Ablation Registration, Segmentation, and Fusion Tool

    PubMed Central

    McCreedy, Evan S.; Cheng, Ruida; Hemler, Paul F.; Viswanathan, Anand; Wood, Bradford J.; McAuliffe, Matthew J.

    2008-01-01

    The Radio Frequency Ablation Segmentation Tool (RFAST) is a software application developed using NIH's Medical Image Processing Analysis and Visualization (MIPAV) API for the specific purpose of assisting physicians in the planning of radio frequency ablation (RFA) procedures. The RFAST application sequentially leads the physician through the steps necessary to register, fuse, segment, visualize and plan the RFA treatment. Three-dimensional volume visualization of the CT dataset with segmented 3D surface models enables the physician to interactively position the ablation probe to simulate burns and to semi-manually simulate sphere packing in an attempt to optimize probe placement. PMID:16871716

  1. The Impact of an Online Crowdsourcing Diagnostic Tool on Health Care Utilization: A Case Study Using a Novel Approach to Retrospective Claims Analysis.

    PubMed

    Juusola, Jessie L; Quisel, Thomas R; Foschini, Luca; Ladapo, Joseph A

    2016-06-01

    Patients with difficult medical cases often remain undiagnosed despite visiting multiple physicians. A new online platform, CrowdMed, uses crowdsourcing to quickly and efficiently reach an accurate diagnosis for these patients. This study sought to evaluate whether CrowdMed decreased health care utilization for patients who have used the service. Novel, electronic methods of patient recruitment and data collection were utilized. Patients who completed cases on CrowdMed's platform between July 2014 and April 2015 were recruited for the study via email and screened via an online survey. After providing eConsent, participants provided identifying information used to access their medical claims data, which was retrieved through a third-party web application program interface (API). Utilization metrics including frequency of provider visits and medical charges were compared pre- and post-case resolution to assess the impact of resolving a case on CrowdMed. Of 45 CrowdMed users who completed the study survey, comprehensive claims data was available via API for 13 participants, who made up the final enrolled sample. There were a total of 221 health care provider visits collected for the study participants, with service dates ranging from September 2013 to July 2015. Frequency of provider visits was significantly lower after resolution of a case on CrowdMed (mean of 1.07 visits per month pre-resolution vs. 0.65 visits per month post-resolution, P=.01). Medical charges were also significantly lower after case resolution (mean of US $719.70 per month pre-resolution vs. US $516.79 per month post-resolution, P=.03). There was no significant relationship between study results and disease onset date, and there was no evidence of regression to the mean influencing results. This study employed technology-enabled methods to demonstrate that patients who used CrowdMed had lower health care utilization after case resolution. However, since the final sample size was limited, results should be interpreted as a case study. Despite this limitation, the statistically significant results suggest that online crowdsourcing shows promise as an efficient method of solving difficult medical cases.

  2. Monitoring of IaaS and scientific applications on the Cloud using the Elasticsearch ecosystem

    NASA Astrophysics Data System (ADS)

    Bagnasco, S.; Berzano, D.; Guarise, A.; Lusso, S.; Masera, M.; Vallero, S.

    2015-05-01

    The private Cloud at the Torino INFN computing centre offers IaaS services to different scientific computing applications. The infrastructure is managed with the OpenNebula cloud controller. The main stakeholders of the facility are a grid Tier-2 site for the ALICE collaboration at LHC, an interactive analysis facility for the same experiment and a grid Tier-2 site for the BES-III collaboration, plus an increasing number of other small tenants. Besides keeping track of the usage, the automation of dynamic allocation of resources to tenants requires detailed monitoring and accounting of the resource usage. As a first investigation towards this, we set up a monitoring system to inspect the site activities both in terms of IaaS and applications running on the hosted virtual instances. For this purpose we used the Elasticsearch, Logstash and Kibana stack. In the current implementation, the heterogeneous accounting information is fed to different MySQL databases and sent to Elasticsearch via a custom Logstash plugin. For the IaaS metering, we developed sensors for the OpenNebula API. The IaaS level information gathered through the API is sent to the MySQL database through an ad-hoc developed RESTful web service, which is also used for other accounting purposes. Concerning the application level, we used the Root plugin TProofMonSenderSQL to collect accounting data from the interactive analysis facility. The BES-III virtual instances used to be monitored with Zabbix, as a proof of concept we also retrieve the information contained in the Zabbix database. Each of these three cases is indexed separately in Elasticsearch. We are now starting to consider dismissing the intermediate level provided by the SQL database and evaluating a NoSQL option as a unique central database for all the monitoring information. We setup a set of Kibana dashboards with pre-defined queries in order to monitor the relevant information in each case. In this way we have achieved a uniform monitoring interface for both the IaaS and the scientific applications, mostly leveraging off-the-shelf tools.

  3. Chimeras of Bet v 1 and Api g 1 reveal heterogeneous IgE responses in patients with birch pollen allergy

    PubMed Central

    Gepp, Barbara; Lengger, Nina; Bublin, Merima; Hemmer, Wolfgang; Breiteneder, Heimo; Radauer, Christian

    2014-01-01

    Background Characterization of IgE-binding epitopes of allergens and determination of their patient-specific relevance is crucial for the diagnosis and treatment of allergy. Objective We sought to assess the contribution of specific surface areas of the major birch pollen allergen Bet v 1.0101 to binding IgE of individual patients. Methods Four distinct areas of Bet v 1 representing in total 81% of its surface were grafted onto the scaffold of its homolog, Api g 1.0101, to yield the chimeras Api-Bet-1 to Api-Bet-4. The chimeras were expressed in Escherichia coli and purified. IgE binding of 64 sera from Bet v 1–sensitized subjects with birch pollen allergy was determined by using direct ELISA. Specificity was assessed by means of inhibition ELISA. Results rApi g 1.0101, Api-Bet-1, Api-Bet-2, Api-Bet-3, and Api-Bet-4 bound IgE from 44%, 89%, 80%, 78%, and 48% of the patients, respectively. By comparing the amount of IgE binding to the chimeras and to rApi g 1.0101, 81%, 70%, 75%, and 45% of the patients showed significantly enhanced IgE binding to Api-Bet-1, Api-Bet-2, Api-Bet-3, and Api-Bet-4, respectively. The minority (8%) of the sera revealed enhanced IgE binding exclusively to a single chimera, whereas 31% showed increased IgE binding to all 4 chimeras compared with rApi g 1.0101. The chimeras inhibited up to 70% of IgE binding to rBet v 1.0101, confirming the specific IgE recognition of the grafted regions. Conclusion The Bet v 1–specific IgE response is polyclonal, and epitopes are spread across the entire Bet v 1 surface. Furthermore, the IgE recognition profile of Bet v 1 is highly patient specific. PMID:24529686

  4. Predominant Api m 10 sensitization as risk factor for treatment failure in honey bee venom immunotherapy.

    PubMed

    Frick, Marcel; Fischer, Jörg; Helbling, Arthur; Ruëff, Franziska; Wieczorek, Dorothea; Ollert, Markus; Pfützner, Wolfgang; Müller, Sabine; Huss-Marp, Johannes; Dorn, Britta; Biedermann, Tilo; Lidholm, Jonas; Ruecker, Gerta; Bantleon, Frank; Miehe, Michaela; Spillner, Edzard; Jakob, Thilo

    2016-12-01

    Component resolution recently identified distinct sensitization profiles in honey bee venom (HBV) allergy, some of which were dominated by specific IgE to Api m 3 and/or Api m 10, which have been reported to be underrepresented in therapeutic HBV preparations. We performed a retrospective analysis of component-resolved sensitization profiles in HBV-allergic patients and association with treatment outcome. HBV-allergic patients who had undergone controlled honey bee sting challenge after at least 6 months of HBV immunotherapy (n = 115) were included and classified as responder (n = 79) or treatment failure (n = 36) on the basis of absence or presence of systemic allergic reactions upon sting challenge. IgE reactivity to a panel of HBV allergens was analyzed in sera obtained before immunotherapy and before sting challenge. No differences were observed between responders and nonresponders regarding levels of IgE sensitization to Api m 1, Api m 2, Api m 3, and Api m 5. In contrast, Api m 10 specific IgE was moderately but significantly increased in nonresponders. Predominant Api m 10 sensitization (>50% of specific IgE to HBV) was the best discriminator (specificity, 95%; sensitivity, 25%) with an odds ratio of 8.444 (2.127-33.53; P = .0013) for treatment failure. Some but not all therapeutic HBV preparations displayed a lack of Api m 10, whereas Api m 1 and Api m 3 immunoreactivity was comparable to that of crude HBV. In line with this, significant Api m 10 sIgG 4 induction was observed only in those patients who were treated with HBV in which Api m 10 was detectable. Component-resolved sensitization profiles in HBV allergy suggest predominant IgE sensitization to Api m 10 as a risk factor for treatment failure in HBV immunotherapy. Copyright © 2016 The Authors. Published by Elsevier Inc. All rights reserved.

  5. Chimeras of Bet v 1 and Api g 1 reveal heterogeneous IgE responses in patients with birch pollen allergy.

    PubMed

    Gepp, Barbara; Lengger, Nina; Bublin, Merima; Hemmer, Wolfgang; Breiteneder, Heimo; Radauer, Christian

    2014-07-01

    Characterization of IgE-binding epitopes of allergens and determination of their patient-specific relevance is crucial for the diagnosis and treatment of allergy. We sought to assess the contribution of specific surface areas of the major birch pollen allergen Bet v 1.0101 to binding IgE of individual patients. Four distinct areas of Bet v 1 representing in total 81% of its surface were grafted onto the scaffold of its homolog, Api g 1.0101, to yield the chimeras Api-Bet-1 to Api-Bet-4. The chimeras were expressed in Escherichia coli and purified. IgE binding of 64 sera from Bet v 1-sensitized subjects with birch pollen allergy was determined by using direct ELISA. Specificity was assessed by means of inhibition ELISA. rApi g 1.0101, Api-Bet-1, Api-Bet-2, Api-Bet-3, and Api-Bet-4 bound IgE from 44%, 89%, 80%, 78%, and 48% of the patients, respectively. By comparing the amount of IgE binding to the chimeras and to rApi g 1.0101, 81%, 70%, 75%, and 45% of the patients showed significantly enhanced IgE binding to Api-Bet-1, Api-Bet-2, Api-Bet-3, and Api-Bet-4, respectively. The minority (8%) of the sera revealed enhanced IgE binding exclusively to a single chimera, whereas 31% showed increased IgE binding to all 4 chimeras compared with rApi g 1.0101. The chimeras inhibited up to 70% of IgE binding to rBet v 1.0101, confirming the specific IgE recognition of the grafted regions. The Bet v 1-specific IgE response is polyclonal, and epitopes are spread across the entire Bet v 1 surface. Furthermore, the IgE recognition profile of Bet v 1 is highly patient specific. Copyright © 2014 The Authors. Published by Mosby, Inc. All rights reserved.

  6. A Web Portal-Based Time-Aware KML Animation Tool for Exploring Spatiotemporal Dynamics of Hydrological Events

    NASA Astrophysics Data System (ADS)

    Bao, X.; Cai, X.; Liu, Y.

    2009-12-01

    Understanding spatiotemporal dynamics of hydrological events such as storms and droughts is highly valuable for decision making on disaster mitigation and recovery. Virtual Globe-based technologies such as Google Earth and Open Geospatial Consortium KML standards show great promises for collaborative exploration of such events using visual analytical approaches. However, currently there are two barriers for wider usage of such approaches. First, there lacks an easy way to use open source tools to convert legacy or existing data formats such as shapefiles, geotiff, or web services-based data sources to KML and to produce time-aware KML files. Second, an integrated web portal-based time-aware animation tool is currently not available. Thus users usually share their files in the portal but have no means to visually explore them without leaving the portal environment which the users are familiar with. We develop a web portal-based time-aware KML animation tool for viewing extreme hydrologic events. The tool is based on Google Earth JavaScript API and Java Portlet standard 2.0 JSR-286, and it is currently deployable in one of the most popular open source portal frameworks, namely Liferay. We have also developed an open source toolkit kml-soc-ncsa (http://code.google.com/p/kml-soc-ncsa/) to facilitate the conversion of multiple formats into KML and the creation of time-aware KML files. We illustrate our tool using some example cases, in which drought and storm events with both time and space dimension can be explored in this web-based KML animation portlet. The tool provides an easy-to-use web browser-based portal environment for multiple users to collaboratively share and explore their time-aware KML files as well as improving the understanding of the spatiotemporal dynamics of the hydrological events.

  7. Temporal and spatial behavior of pharmaceuticals in ...

    EPA Pesticide Factsheets

    The behavior of active pharmaceutical ingredients (APIs) in urban estuaries is not well understood. In this study, 15 high volume usage APIs were measured over a one year period throughout Narragansett Bay, RI, USA to determine factors controlling their concentration and distribution. Dissolved APIs ranged in concentration from not detected to 310 ng/L, with numerous APIs present at all sites and sampling periods. Eight APIs were present in suspended particulate material, ranging in concentration from <1 ng/g to 44 ng/g. Partitioning coefficients (Kds) were determined for APIs present in both the dissolved and particulate phases, with their range and variability remaining relatively constant during the study. Organic carbon normalization reduced the observed variability of several APIs to a small extent; however, other factors appear to play a role in controlling partitioning behavior. The continuous discharge of wastewater treatment plant effluents into upper Narragansett Bay resulted in sustained levels of APIs, resulting in a zone of “pseudo-persistence.” For most of the APIs, there was a strong relationship with salinity, indicating conservative behavior within the estuary. Short flushing times in Narragansett Bay coupled with APIs present primarily in the dissolved phase suggests that most APIs will be diluted and transported out of the estuary, with only small amounts of several compounds removed to and sequestered in sediments. This study ide

  8. 49 CFR 195.565 - How do I install cathodic protection on breakout tanks?

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ...) capacity built to API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the system in accordance with API Recommended Practice 651. However, installation of the system need not comply with API Recommended Practice 651 on any tank for which you note in...

  9. 49 CFR 195.579 - What must I do to mitigate internal corrosion?

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the lining in accordance with API Recommended Practice 652. However, installation of the lining need not comply with API Recommended Practice 652 on any tank for which you note in the corrosion...

  10. 49 CFR 195.565 - How do I install cathodic protection on breakout tanks?

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ...) capacity built to API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the system in accordance with API Recommended Practice 651. However, installation of the system need not comply with API Recommended Practice 651 on any tank for which you note in...

  11. 49 CFR 195.565 - How do I install cathodic protection on breakout tanks?

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ...) capacity built to API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the system in accordance with API Recommended Practice 651. However, installation of the system need not comply with API Recommended Practice 651 on any tank for which you note in...

  12. 49 CFR 195.579 - What must I do to mitigate internal corrosion?

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the lining in accordance with API Recommended Practice 652. However, installation of the lining need not comply with API Recommended Practice 652 on any tank for which you note in the corrosion...

  13. 49 CFR 195.579 - What must I do to mitigate internal corrosion?

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the lining in accordance with API Recommended Practice 652. However, installation of the lining need not comply with API Recommended Practice 652 on any tank for which you note in the corrosion...

  14. 49 CFR 195.579 - What must I do to mitigate internal corrosion?

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the lining in accordance with API Recommended Practice 652. However, installation of the lining need not comply with API Recommended Practice 652 on any tank for which you note in the corrosion...

  15. 78 FR 48738 - Self-Regulatory Organizations; C2 Options Exchange, Incorporated; Notice of Filing and Immediate...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-08-09

    ... depend upon the Application Programming Interface (``API'') a Permit Holder is using.\\4\\ Currently, the Exchange offers two APIs: CBOE Market Interface (``CMi'') API and Financial Information eXchange (``FIX... available APIs, and if applicable, which version, it would like to use. \\4\\ An API is a computer interface...

  16. 49 CFR 195.565 - How do I install cathodic protection on breakout tanks?

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ...) capacity built to API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the system in accordance with API Recommended Practice 651. However, installation of the system need not comply with API Recommended Practice 651 on any tank for which you note in...

  17. 49 CFR 195.579 - What must I do to mitigate internal corrosion?

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the lining in accordance with API Recommended Practice 652. However, installation of the lining need not comply with API Recommended Practice 652 on any tank for which you note in the corrosion...

  18. 49 CFR 195.565 - How do I install cathodic protection on breakout tanks?

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ...) capacity built to API Specification 12F, API Standard 620, or API Standard 650 (or its predecessor Standard 12C), you must install the system in accordance with API Recommended Practice 651. However, installation of the system need not comply with API Recommended Practice 651 on any tank for which you note in...

  19. Judicious use of custom development in an open source component architecture

    NASA Astrophysics Data System (ADS)

    Bristol, S.; Latysh, N.; Long, D.; Tekell, S.; Allen, J.

    2014-12-01

    Modern software engineering is not as much programming from scratch as innovative assembly of existing components. Seamlessly integrating disparate components into scalable, performant architecture requires sound engineering craftsmanship and can often result in increased cost efficiency and accelerated capabilities if software teams focus their creativity on the edges of the problem space. ScienceBase is part of the U.S. Geological Survey scientific cyberinfrastructure, providing data and information management, distribution services, and analysis capabilities in a way that strives to follow this pattern. ScienceBase leverages open source NoSQL and relational databases, search indexing technology, spatial service engines, numerous libraries, and one proprietary but necessary software component in its architecture. The primary engineering focus is cohesive component interaction, including construction of a seamless Application Programming Interface (API) across all elements. The API allows researchers and software developers alike to leverage the infrastructure in unique, creative ways. Scaling the ScienceBase architecture and core API with increasing data volume (more databases) and complexity (integrated science problems) is a primary challenge addressed by judicious use of custom development in the component architecture. Other data management and informatics activities in the earth sciences have independently resolved to a similar design of reusing and building upon established technology and are working through similar issues for managing and developing information (e.g., U.S. Geoscience Information Network; NASA's Earth Observing System Clearing House; GSToRE at the University of New Mexico). Recent discussions facilitated through the Earth Science Information Partners are exploring potential avenues to exploit the implicit relationships between similar projects for explicit gains in our ability to more rapidly advance global scientific cyberinfrastructure.

  20. AWS-Glacier As A Storage Foundation For AWS-EC2 Hosted Scientific Data Services

    NASA Astrophysics Data System (ADS)

    Gallagher, J. H. R.; Potter, N.

    2016-12-01

    Using AWS Glacier as a base level data store for a scientific data service presents new challenges for the web accessible data services, along with their software clients and human operators. All meaningful Glacier transactions take at least 4 hours to complete. This is in contrast to the various web APIs for data such as WMS, WFS, WCS, DAP2, and Netcdf tools which were all written based on the premise that the response will be (nearly) immediate. Only DAP4 and WPS contain an explicit asynchronous component to their respective protocols which allows for "return later" behaviors. We were able to put Hyrax (a DAP4 server) in front of Glacier-held resources, but there were significant issues. Any kind of probing of the datasets happens at the cost of the Glacier retrieval period, 4 hours. A couple of crucial things fall out of this: The first is that the service must cache metadata, including coordinate map arrays, so that a client can have enough information available in the "immediate" time frame to make a decisions about what to ask for from the dataset. This type of request planning is important because a data access request will take 4 hours to complete unless the data resource has been cached. The second thing is that the clients need to change their behavior when accessing datasets in an asynchronous system, even if the metadata is cached. Commonly, client applications will request a number of data components from a DAP2 service in the course of "discovering" the dataset. This may not be a well-supported model of interaction with Glacier or any other high latency data store.

  1. An application programming interface for extreme precipitation and hazard products

    NASA Astrophysics Data System (ADS)

    Kirschbaum, D.; Stanley, T.; Cappelaere, P. G.; Reed, J.; Lammers, M.

    2016-12-01

    Remote sensing data provides situational awareness of extreme events and hazards over large areas in a way that is impossible to achieve with in situ data. However, more valuable than raw data is actionable information based on user needs. This information can take the form of derived products, extraction of a subset of variables in a larger data matrix, or data processing for a specific goal. These products can then stream to the end users, who can use these data to improve local to global decision making. This presentation will outline both the science and methodology of two new data products and tools that can provide relevant climate and hazard data for response and support. The Global Precipitation Measurement (GPM) mission provides near real-time information on rain and snow around the world every thirty minutes. Through a new applications programing interface (API), this data can be freely accessed by consumers to visualize, analyze, and communicate where, when and how much rain is falling worldwide. The second tool is a global landslide model that provides situational awareness of potential landslide activity in near real-time, utilizing several remotely sensed data products. This hazard information is also provided through an API and is being ingested by the emergency response community, international aid organizations, and others around the world. This presentation will highlight lessons learned through the development, implementation, and communication of these products and tools with the goal of enabling better and more effective decision making.

  2. Reduced Crystallization Temperature Methodology for Polymer Selection in Amorphous Solid Dispersions: Stability Perspective.

    PubMed

    Bhugra, Chandan; Telang, Chitra; Schwabe, Robert; Zhong, Li

    2016-09-06

    API-polymer interactions, used to select the right polymeric matrix with an aim to stabilize an amorphous dispersion, are routinely studied using spectroscopic and/or calorimetric techniques (i.e., melting point depression). An alternate selection tool has been explored to rank order polymers for formation of stable amorphous dispersions as a pragmatic method for polymer selection. Reduced crystallization temperature of API, a parameter introduced by Zhou et al.,1 was utilized in this study for rank ordering interactions in API-polymeric systems. The trends in reduced crystallization temperature monitored over polymer concentration range of up to 20% polymer loading were utilized to calculate "crystallization parameter" or CP for two model systems (nifedipine and BI ABC). The rank order of CP, i.e., a measure of API-polymer interaction, for nifedipine followed the order PVP > PVP-VA > Soluplus > HPMCAS > PV Ac > PAA. This rank ordering was correlated to published results of molecular interactions and physical stability for nifedipine. A different rank ordering was observed for BI ABC: PAA > PVP > HPMCAS > Soluplus > PVPV-VA > PVAc. Interactions for BI ABC were not as differentiated when compared to nifedipine based on CP trends. BI ABC dispersions at drug loadings between 40 and 60% were physically stable for prolonged periods under ICH conditions as well as accelerated stress. We propose that large CP differences among polymers could be predictive of stability outcomes. Acceptable stability at pharmaceutically relevant drug loadings would suggest that the relative influence of downstream processes, such as polymer solubility in various solvents, process suitability and selection, and more importantly supersaturation potential, should be higher compared to stability considerations while developing compounds like BI ABC.

  3. 49 CFR 194.105 - Worst case discharge.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ...: Prevention measure Standard Credit(percent) Secondary containment >100% NFPA 30 50 Built/repaired to API standards API STD 620/650/653 10 Overfill protection standards API RP 2350 5 Testing/cathodic protection API...

  4. 49 CFR 194.105 - Worst case discharge.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ...: Prevention measure Standard Credit(percent) Secondary containment > 100% NFPA 30 50 Built/repaired to API standards API STD 620/650/653 10 Overfill protection standards API RP 2350 5 Testing/cathodic protection API...

  5. 49 CFR 194.105 - Worst case discharge.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ...: Prevention measure Standard Credit(percent) Secondary containment >100% NFPA 30 50 Built/repaired to API standards API STD 620/650/653 10 Overfill protection standards API RP 2350 5 Testing/cathodic protection API...

  6. 49 CFR 194.105 - Worst case discharge.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ...: Prevention measure Standard Credit(percent) Secondary containment > 100% NFPA 30 50 Built/repaired to API standards API STD 620/650/653 10 Overfill protection standards API RP 2350 5 Testing/cathodic protection API...

  7. 49 CFR 194.105 - Worst case discharge.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ...: Prevention measure Standard Credit(percent) Secondary containment > 100% NFPA 30 50 Built/repaired to API standards API STD 620/650/653 10 Overfill protection standards API RP 2350 5 Testing/cathodic protection API...

  8. The novel Akt inhibitor API-1 induces c-FLIP degradation and synergizes with TRAIL to augment apoptosis independent of Akt inhibition

    PubMed Central

    Li, Bo; Ren, Hui; Yue, Ping; Chen, Mingwei; Khuri, Fadlo R.; Sun, Shi-Yong

    2012-01-01

    API-1 is a novel small molecule inhibitor of Akt, which acts by binding to Akt and preventing its membrane translocation, and has promising preclinical antitumor activity. In this study, we reveal a novel function of API-1 in regulation of c-FLIP levels and tumor necrosis factor-related apoptosis-inducing ligand (TRAIL)-induced apoptosis, independent of Akt inhibition. API-1 effectively induced apoptosis in tested cancer cell lines including activation of caspase-8 and caspase-9. It reduced the levels of c-FLIP without increasing the expression of DR4 or DR5. Accordingly, it synergized with TRAIL to induce apoptosis. Enforced expression of ectopic c-FLIP did not attenuate API-1-induced apoptosis, but inhibited its ability to enhance TRAIL-induced apoptosis. These data indicate that downregulation of c-FLIP mediates enhancement of TRAIL-induced apoptosis by API-1, but is not sufficient for API-1-induced apoptosis. API-1-induced reduction of c-FLIP could be blocked by the proteasome inhibitor MG132. Moreover, API-1 increased c-FLIP ubiquitination and decreased c-FLIP stability. These data together suggest that API-1 downregulates c-FLIP by facilitating its ubiquitination and proteasome-mediated degradation. Since other Akt inhibitors including API-2 and MK2206 had minimal effects on reducing c-FLIP and enhancement of TRAIL-induced apoptosis, it is likely that API-1 reduces c-FLIP and enhances TRAIL-induced apoptosis independent of its Akt-inhibitory activity. PMID:22345097

  9. A Ruby API to query the Ensembl database for genomic features.

    PubMed

    Strozzi, Francesco; Aerts, Jan

    2011-04-01

    The Ensembl database makes genomic features available via its Genome Browser. It is also possible to access the underlying data through a Perl API for advanced querying. We have developed a full-featured Ruby API to the Ensembl databases, providing the same functionality as the Perl interface with additional features. A single Ruby API is used to access different releases of the Ensembl databases and is also able to query multi-species databases. Most functionality of the API is provided using the ActiveRecord pattern. The library depends on introspection to make it release independent. The API is available through the Rubygem system and can be installed with the command gem install ruby-ensembl-api.

  10. Grid workflow job execution service 'Pilot'

    NASA Astrophysics Data System (ADS)

    Shamardin, Lev; Kryukov, Alexander; Demichev, Andrey; Ilyin, Vyacheslav

    2011-12-01

    'Pilot' is a grid job execution service for workflow jobs. The main goal for the service is to automate computations with multiple stages since they can be expressed as simple workflows. Each job is a directed acyclic graph of tasks and each task is an execution of something on a grid resource (or 'computing element'). Tasks may be submitted to any WS-GRAM (Globus Toolkit 4) service. The target resources for the tasks execution are selected by the Pilot service from the set of available resources which match the specific requirements from the task and/or job definition. Some simple conditional execution logic is also provided. The 'Pilot' service is built on the REST concepts and provides a simple API through authenticated HTTPS. This service is deployed and used in production in a Russian national grid project GridNNN.

  11. Infra-Population and -Community Dynamics of the Parasites Nosema apis and Nosema ceranae, and Consequences for Honey Bee (Apis mellifera) Hosts

    PubMed Central

    Williams, Geoffrey R.; Shutler, Dave; Burgher-MacLellan, Karen L.; Rogers, Richard E. L.

    2014-01-01

    Nosema spp. fungal gut parasites are among myriad possible explanations for contemporary increased mortality of western honey bees (Apis mellifera, hereafter honey bee) in many regions of the world. Invasive Nosema ceranae is particularly worrisome because some evidence suggests it has greater virulence than its congener N. apis. N. ceranae appears to have recently switched hosts from Asian honey bees (Apis cerana) and now has a nearly global distribution in honey bees, apparently displacing N. apis. We examined parasite reproduction and effects of N. apis, N. ceranae, and mixed Nosema infections on honey bee hosts in laboratory experiments. Both infection intensity and honey bee mortality were significantly greater for N. ceranae than for N. apis or mixed infections; mixed infection resulted in mortality similar to N. apis parasitism and reduced spore intensity, possibly due to inter-specific competition. This is the first long-term laboratory study to demonstrate lethal consequences of N. apis and N. ceranae and mixed Nosema parasitism in honey bees, and suggests that differences in reproduction and intra-host competition may explain apparent heterogeneous exclusion of the historic parasite by the invasive species. PMID:24987989

  12. Infra-population and -community dynamics of the parasites Nosema apis and Nosema ceranae, and consequences for honey bee (Apis mellifera) hosts.

    PubMed

    Williams, Geoffrey R; Shutler, Dave; Burgher-MacLellan, Karen L; Rogers, Richard E L

    2014-01-01

    Nosema spp. fungal gut parasites are among myriad possible explanations for contemporary increased mortality of western honey bees (Apis mellifera, hereafter honey bee) in many regions of the world. Invasive Nosema ceranae is particularly worrisome because some evidence suggests it has greater virulence than its congener N. apis. N. ceranae appears to have recently switched hosts from Asian honey bees (Apis cerana) and now has a nearly global distribution in honey bees, apparently displacing N. apis. We examined parasite reproduction and effects of N. apis, N. ceranae, and mixed Nosema infections on honey bee hosts in laboratory experiments. Both infection intensity and honey bee mortality were significantly greater for N. ceranae than for N. apis or mixed infections; mixed infection resulted in mortality similar to N. apis parasitism and reduced spore intensity, possibly due to inter-specific competition. This is the first long-term laboratory study to demonstrate lethal consequences of N. apis and N. ceranae and mixed Nosema parasitism in honey bees, and suggests that differences in reproduction and intra-host competition may explain apparent heterogeneous exclusion of the historic parasite by the invasive species.

  13. Transcriptome differences in the hypopharyngeal gland between Western Honeybees (Apis mellifera) and Eastern Honeybees (Apis cerana).

    PubMed

    Liu, Hao; Wang, Zi-Long; Tian, Liu-Qing; Qin, Qiu-Hong; Wu, Xiao-Bo; Yan, Wei-Yu; Zeng, Zhi-Jiang

    2014-08-30

    Apis mellifera and Apis cerana are two sibling species of Apidae. Apis cerana is adept at collecting sporadic nectar in mountain and forest region and exhibits stiffer hardiness and acarid resistance as a result of natural selection, whereas Apis mellifera has the advantage of producing royal jelly. To identify differentially expressed genes (DEGs) that affect the development of hypopharyngeal gland (HG) and/or the secretion of royal jelly between these two honeybee species, we performed a digital gene expression (DGE) analysis of the HGs of these two species at three developmental stages (newly emerged worker, nurse and forager). Twelve DGE-tag libraries were constructed and sequenced using the total RNA extracted from the HGs of newly emerged workers, nurses, and foragers of Apis mellifera and Apis cerana. Finally, a total of 1482 genes in Apis mellifera and 1313 in Apis cerana were found to exhibit an expression difference among the three developmental stages. A total of 1417 DEGs were identified between these two species. Of these, 623, 1072, and 462 genes showed an expression difference at the newly emerged worker, nurse, and forager stages, respectively. The nurse stage exhibited the highest number of DEGs between these two species and most of these were found to be up-regulated in Apis mellifera. These results suggest that the higher yield of royal jelly in Apis mellifera may be due to the higher expression level of these DEGs. In this study, we investigated the DEGs between the HGs of two sibling honeybee species (Apis mellifera and Apis cerana). Our results indicated that the gene expression difference was associated with the difference in the royal jelly yield between these two species. These results provide an important clue for clarifying the mechanisms underlying hypopharyngeal gland development and the production of royal jelly.

  14. Detecting Potentially Compromised Credentials in a Large-Scale Production Single-Signon System

    DTIC Science & Technology

    2014-06-01

    Attention Deficit Hyperactivity Disorder ( ADHD ), Post-Traumatic Stress Disorder (PTSD), anxiety, they are neurotic, and have memory issues. They... Deficit Hyperactivity Disorder API Application Programming Interface CAC Common Access Card CBL Composite Blocking List CDF Cumulative Distribution...Service Logons (DSLs) system . . . . . . . . . . . . . . . . 49 xi THIS PAGE INTENTIONALLY LEFT BLANK xii List of Acronyms and Abbreviations ADHD Attention

  15. 49 CFR 195.264 - Impoundment, protection against entry, normal/emergency venting or pressure/vacuum relief for...

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... specified: (1) For tanks built to API Specification 12F, API Standard 620, and others (such as API Standard... accordance with section 4.3.2.3.1. (2) For tanks built to API 2510, the installation of impoundment must be in accordance with section 5 or 11 of API 2510 (incorporated by reference, see § 195.3). (c...

  16. 49 CFR 195.264 - Impoundment, protection against entry, normal/emergency venting or pressure/vacuum relief for...

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... specified: (1) For tanks built to API Specification 12F, API Standard 620, and others (such as API Standard... accordance with section 4.3.2.3.1. (2) For tanks built to API 2510, the installation of impoundment must be in accordance with section 5 or 11 of API 2510 (incorporated by reference, see § 195.3). (c...

  17. 49 CFR 195.264 - Impoundment, protection against entry, normal/emergency venting or pressure/vacuum relief for...

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... specified: (1) For tanks built to API Specification 12F, API Standard 620, and others (such as API Standard... accordance with section 4.3.2.3.1. (2) For tanks built to API 2510, the installation of impoundment must be in accordance with section 5 or 11 of API 2510 (incorporated by reference, see § 195.3). (c...

  18. 49 CFR 195.264 - Impoundment, protection against entry, normal/emergency venting or pressure/vacuum relief for...

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... specified: (1) For tanks built to API Specification 12F, API Standard 620, and others (such as API Standard... accordance with section 4.3.2.3.1. (2) For tanks built to API 2510, the installation of impoundment must be in accordance with section 5 or 11 of API 2510 (incorporated by reference, see § 195.3). (c...

  19. Predicting concentrations of trace organic compounds in municipal wastewater treatment plant sludge and biosolids using the PhATE™ model.

    PubMed

    Cunningham, Virginia L; D'Aco, Vincent J; Pfeiffer, Danielle; Anderson, Paul D; Buzby, Mary E; Hannah, Robert E; Jahnke, James; Parke, Neil J

    2012-07-01

    This article presents the capability expansion of the PhATE™ (pharmaceutical assessment and transport evaluation) model to predict concentrations of trace organics in sludges and biosolids from municipal wastewater treatment plants (WWTPs). PhATE was originally developed as an empirical model to estimate potential concentrations of active pharmaceutical ingredients (APIs) in US surface and drinking waters that could result from patient use of medicines. However, many compounds, including pharmaceuticals, are not completely transformed in WWTPs and remain in biosolids that may be applied to land as a soil amendment. This practice leads to concerns about potential exposures of people who may come into contact with amended soils and also about potential effects to plants and animals living in or contacting such soils. The model estimates the mass of API in WWTP influent based on the population served, the API per capita use, and the potential loss of the compound associated with human use (e.g., metabolism). The mass of API on the treated biosolids is then estimated based on partitioning to primary and secondary solids, potential loss due to biodegradation in secondary treatment (e.g., activated sludge), and potential loss during sludge treatment (e.g., aerobic digestion, anaerobic digestion, composting). Simulations using 2 surrogate compounds show that predicted environmental concentrations (PECs) generated by PhATE are in very good agreement with measured concentrations, i.e., well within 1 order of magnitude. Model simulations were then carried out for 18 APIs representing a broad range of chemical and use characteristics. These simulations yielded 4 categories of results: 1) PECs are in good agreement with measured data for 9 compounds with high analytical detection frequencies, 2) PECs are greater than measured data for 3 compounds with high analytical detection frequencies, possibly as a result of as yet unidentified depletion mechanisms, 3) PECs are less than analytical reporting limits for 5 compounds with low analytical detection frequencies, and 4) the PEC is greater than the analytical method reporting limit for 1 compound with a low analytical detection frequency, possibly again as a result of insufficient depletion data. Overall, these results demonstrate that PhATE has the potential to be a very useful tool in the evaluation of APIs in biosolids. Possible applications include: prioritizing APIs for assessment even in the absence of analytical methods; evaluating sludge processing scenarios to explore potential mitigation approaches; using in risk assessments; and developing realistic nationwide concentrations, because PECs can be represented as a cumulative probability distribution. Finally, comparison of PECs to measured concentrations can also be used to identify the need for fate studies of compounds of interest in biosolids. Copyright © 2011 SETAC.

  20. The novel Akt inhibitor API-1 induces c-FLIP degradation and synergizes with TRAIL to augment apoptosis independent of Akt inhibition.

    PubMed

    Li, Bo; Ren, Hui; Yue, Ping; Chen, Mingwei; Khuri, Fadlo R; Sun, Shi-Yong

    2012-04-01

    API-1 (pyrido[2,3-d]pyrimidines) is a novel small-molecule inhibitor of Akt, which acts by binding to Akt and preventing its membrane translocation and has promising preclinical antitumor activity. In this study, we reveal a novel function of API-1 in regulation of cellular FLICE-inhibitory protein (c-FLIP) levels and TRAIL-induced apoptosis, independent of Akt inhibition. API-1 effectively induced apoptosis in tested cancer cell lines including activation of caspase-8 and caspase-9. It reduced the levels of c-FLIP without increasing the expression of death receptor 4 (DR4) or DR5. Accordingly, it synergized with TRAIL to induce apoptosis. Enforced expression of ectopic c-FLIP did not attenuate API-1-induced apoptosis but inhibited its ability to enhance TRAIL-induced apoptosis. These data indicate that downregulation of c-FLIP mediates enhancement of TRAIL-induced apoptosis by API-1 but is not sufficient for API-1-induced apoptosis. API-1-induced reduction of c-FLIP could be blocked by the proteasome inhibitor MG132. Moreover, API-1 increased c-FLIP ubiquitination and decreased c-FLIP stability. These data together suggest that API-1 downregulates c-FLIP by facilitating its ubiquitination and proteasome-mediated degradation. Because other Akt inhibitors including API-2 and MK2206 had minimal effects on reducing c-FLIP and enhancement of TRAIL-induced apoptosis, it is likely that API-1 reduces c-FLIP and enhances TRAIL-induced apoptosis independent of its Akt-inhibitory activity. 2012 AACR

  1. The Best of All Possible Worlds: Applying the Model Driven Architecture Approach to a JC3IEDM OWL Ontology Modeled in UML

    DTIC Science & Technology

    2014-06-01

    from the ODM standard. Leveraging SPARX EA’s Java application programming interface (API), the team built a tool called OWL2EA that can ingest an OWL...server MySQL creates the physical schema that enables a user to store and retrieve data conforming to the vocabulary of the JC3IEDM. 6. GENERATING AN

  2. Aladin Lite: Lightweight sky atlas for browsers

    NASA Astrophysics Data System (ADS)

    Boch, Thomas

    2014-02-01

    Aladin Lite is a lightweight version of the Aladin tool, running in the browser and geared towards simple visualization of a sky region. It allows visualization of image surveys (JPEG multi-resolution HEALPix all-sky surveys) and permits superimposing tabular (VOTable) and footprints (STC-S) data. Aladin Lite is powered by HTML5 canvas technology and is easily embeddable on any web page and can also be controlled through a Javacript API.

  3. Investigation of polymorphic transitions of piracetam induced during wet granulation.

    PubMed

    Potter, Catherine B; Kollamaram, Gayathri; Zeglinski, Jacek; Whitaker, Darren A; Croker, Denise M; Walker, Gavin M

    2017-10-01

    Piracetam was investigated as a model API which is known to exhibit a number of different polymorphic forms. It is freely soluble in water so the possibility exists for polymorphic transformations to occur during wet granulation. Analysis of the polymorphic form present during lab-scale wet granulation, using water as a granulation liquid, was studied with powder X-ray diffraction and Raman spectroscopy as off-line and inline analysis tools respectively. Different excipients with a range of hydrophilicities, aqueous solubilities and molecular weights were investigated to examine their influence on these solution-mediated polymorphic transitions and experimental results were rationalised using molecular modelling. Our results indicated that as an increasing amount of water was added to the as-received piracetam FIII, a greater amount of the API dissolved which recrystallised upon drying to the metastable FII(6.403) via a monohydrate intermediary. Molecular level analysis revealed that the observed preferential transformation of monohydrate to FII is linked with a greater structural similarity between the monohydrate and FII polymorph in comparison to FIII. The application of Raman spectroscopy as a process analytical technology (PAT) tool to monitor the granulation process for the production of the monohydrate intermediate as a precursor to the undesirable metastable form was demonstrated. Copyright © 2017 Elsevier B.V. All rights reserved.

  4. Enabling Tools and Methods for International, Inter-disciplinary and Educational Collaboration

    NASA Astrophysics Data System (ADS)

    Robinson, E. M.; Hoijarvi, K.; Falke, S.; Fialkowski, E.; Kieffer, M.; Husar, R. B.

    2008-05-01

    In the past, collaboration has taken place in tightly-knit workgroups where the members had direct connections to each other. Such collaboration was confined to small workgroups and person-to-person communication. Recent developments through the Internet foster virtual workgroups and organizations where dynamic, 'just-in-time' collaboration can take place over a much larger scale. The emergence of virtual workgroups has strongly influenced the interaction of inter-national, inter-disciplinary, as well as educational activities. In this paper we present an array of enabling tools and methods that incorporate the new technologies including web services, software mashups, tag-based structuring and searching, and wikis for collaborative writing and content organization. Large monolithic, 'do-it-all' software tools are giving way to web service modules, combined through service chaining. Application software can now be created using Service Oriented Architecture (SOA). In the air quality community, data providers and users are distributed in space and time creating barriers for data access. By exposing the data on the internet the space, time barriers are lessened. The federated data system, DataFed, developed at Washington University, accesses data from autonomous, distributed providers. Through data "wrappers", DataFed provides uniform and standards-based access services to heterogeneous, distributed data. Service orientation not only lowers the entry resistance for service providers, but it also allows the creation of user-defined applications and/or mashups. For example, Google Earth's open API allowed many groups to mash their content with Google Earth. Ad hoc tagging gives a rich description of the internet resources, but it has the disadvantage of providing a fuzzy schema. The semantic uniformity of the internet resources can be improved by controlled tagging which apply a consistent namespace and tag combinations to diverse objects. One example of this is the photo-sharing web application Flickr. Just like data, by exposing photos through the internet those can be reused in ways unknown and unanticipated by the provider. For air quality application, Flickr allowed a rich collection of images of forest fire smoke, wind blown dust and haze events to be tagged with controlled tags and used in for evaluating subtle features of the events. Wikis, originally used just for collaboratively writing and discuss documents, are now also a social software workflow managers. In air quality data, wikis provides the means to collaboratively create rich metadata. Wikis become a virtual meeting place to discuss ideas before a workshop of conference, display tagged internet resources, and collaboratively work on documents. Wikis are also useful in the classroom. For instance in class projects, the wiki displays harvested resources, maintains collaborative documents and discussions and is the organizational memory for the project.

  5. The Vital Planning and Analysis (ViPA) ORBAT Data Service Architecture and Design Overview

    DTIC Science & Technology

    2016-07-01

    ORBAT Data Service is a  Web  Service for storing force structures to be used by the ADF. The service is composed into what is known   as   a   Service...DRN) and the Defence Secret Network (DSN) by Defence project JP2030. The ORBAT Data Service is a  Web  Service for storing force structures to be used by...the ADF.    Web  Services are software used to share business logic and data across a network through   application  programming   interfaces   (APIs

  6. Disparities in Rates of Inpatient Mortality and Adverse Events: Race/Ethnicity and Language as Independent Contributors

    PubMed Central

    Hines, Anika L.; Andrews, Roxanne M.; Moy, Ernest; Barrett, Marguerite L.; Coffey, Rosanna M.

    2014-01-01

    Patients with limited English proficiency have known limitations accessing health care, but differences in hospital outcomes once access is obtained are unknown. We investigate inpatient mortality rates and obstetric trauma for self-reported speakers of English, Spanish, and languages of Asia and the Pacific Islands (API) and compare quality of care by language with patterns by race/ethnicity. Data were from the United States Agency for Healthcare Research and Quality, Healthcare Cost and Utilization Project, 2009 State Inpatient Databases for California. There were 3,757,218 records. Speaking a non-English principal language and having a non-White race/ethnicity did not place patients at higher risk for inpatient mortality; the exception was significantly higher stroke mortality for Japanese-speaking patients. Patients who spoke API languages or had API race/ethnicity had higher risk for obstetric trauma than English-speaking White patients. Spanish-speaking Hispanic patients had more obstetric trauma than English-speaking Hispanic patients. The influence of language on obstetric trauma and the potential effects of interpretation services on inpatient care are discussed. The broader context of policy implications for collection and reporting of language data is also presented. Results from other countries with and without English as a primary language are needed for the broadest interpretation and generalization of outcomes. PMID:25514153

  7. 16 CFR 311.4 - Testing.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... (“API”), Publication 1509, Thirteenth Edition, January 1995. API Publication 1509, Thirteenth Edition has been updated to API Publication 1509, Fifteenth Edition, April 2002. API Publication 1509... the materials incorporated by reference may be obtained from: API, 1220 L Street, NW., Washington, DC...

  8. 16 CFR 311.4 - Testing.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... (“API”), Publication 1509, Thirteenth Edition, January 1995. API Publication 1509, Thirteenth Edition has been updated to API Publication 1509, Fifteenth Edition, April 2002. API Publication 1509... the materials incorporated by reference may be obtained from: API, 1220 L Street, NW., Washington, DC...

  9. 16 CFR 311.4 - Testing.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... (“API”), Publication 1509, Thirteenth Edition, January 1995. API Publication 1509, Thirteenth Edition has been updated to API Publication 1509, Fifteenth Edition, April 2002. API Publication 1509... the materials incorporated by reference may be obtained from: API, 1220 L Street, NW., Washington, DC...

  10. 16 CFR 311.4 - Testing.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... (“API”), Publication 1509, Thirteenth Edition, January 1995. API Publication 1509, Thirteenth Edition has been updated to API Publication 1509, Fifteenth Edition, April 2002. API Publication 1509... the materials incorporated by reference may be obtained from: API, 1220 L Street, NW., Washington, DC...

  11. Marshall Space Flight Center Telescience Resource Kit

    NASA Technical Reports Server (NTRS)

    Wade, Gina

    2016-01-01

    Telescience Resource Kit (TReK) is a suite of software applications that can be used to monitor and control assets in space or on the ground. The Telescience Resource Kit was originally developed for the International Space Station program. Since then it has been used to support a variety of NASA programs and projects including the WB-57 Ascent Vehicle Experiment (WAVE) project, the Fast Affordable Science and Technology Satellite (FASTSAT) project, and the Constellation Program. The Payloads Operations Center (POC), also known as the Payload Operations Integration Center (POIC), provides the capability for payload users to operate their payloads at their home sites. In this environment, TReK provides local ground support system services and an interface to utilize remote services provided by the POC. TReK provides ground system services for local and remote payload user sites including International Partner sites, Telescience Support Centers, and U.S. Investigator sites in over 40 locations worldwide. General Capabilities: Support for various data interfaces such as User Datagram Protocol, Transmission Control Protocol, and Serial interfaces. Data Services - retrieve, process, record, playback, forward, and display data (ground based data or telemetry data). Command - create, modify, send, and track commands. Command Management - Configure one TReK system to serve as a command server/filter for other TReK systems. Database - databases are used to store telemetry and command definition information. Application Programming Interface (API) - ANSI C interface compatible with commercial products such as Visual C++, Visual Basic, LabVIEW, Borland C++, etc. The TReK API provides a bridge for users to develop software to access and extend TReK services. Environments - development, test, simulations, training, and flight. Includes standalone training simulators.

  12. Making geospatial data in ASF archive readily accessible

    NASA Astrophysics Data System (ADS)

    Gens, R.; Hogenson, K.; Wolf, V. G.; Drew, L.; Stern, T.; Stoner, M.; Shapran, M.

    2015-12-01

    The way geospatial data is searched, managed, processed and used has changed significantly in recent years. A data archive such as the one at the Alaska Satellite Facility (ASF), one of NASA's twelve interlinked Distributed Active Archive Centers (DAACs), used to be searched solely via user interfaces that were specifically developed for its particular archive and data sets. ASF then moved to using an application programming interface (API) that defined a set of routines, protocols, and tools for distributing the geospatial information stored in the database in real time. This provided a more flexible access to the geospatial data. Yet, it was up to user to develop the tools to get a more tailored access to the data they needed. We present two new approaches for serving data to users. In response to the recent Nepal earthquake we developed a data feed for distributing ESA's Sentinel data. Users can subscribe to the data feed and are provided with the relevant metadata the moment a new data set is available for download. The second approach was an Open Geospatial Consortium (OGC) web feature service (WFS). The WFS hosts the metadata along with a direct link from which the data can be downloaded. It uses the open-source GeoServer software (Youngblood and Iacovella, 2013) and provides an interface to include the geospatial information in the archive directly into the user's geographic information system (GIS) as an additional data layer. Both services are run on top of a geospatial PostGIS database, an open-source geographic extension for the PostgreSQL object-relational database (Marquez, 2015). Marquez, A., 2015. PostGIS essentials. Packt Publishing, 198 p. Youngblood, B. and Iacovella, S., 2013. GeoServer Beginner's Guide, Packt Publishing, 350 p.

  13. A Reasoning And Hypothesis-Generation Framework Based On Scalable Graph Analytics

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sukumar, Sreenivas Rangan

    Finding actionable insights from data has always been difficult. As the scale and forms of data increase tremendously, the task of finding value becomes even more challenging. Data scientists at Oak Ridge National Laboratory are leveraging unique leadership infrastructure (e.g. Urika-XA and Urika-GD appliances) to develop scalable algorithms for semantic, logical and statistical reasoning with unstructured Big Data. We present the deployment of such a framework called ORiGAMI (Oak Ridge Graph Analytics for Medical Innovations) on the National Library of Medicine s SEMANTIC Medline (archive of medical knowledge since 1994). Medline contains over 70 million knowledge nuggets published in 23.5more » million papers in medical literature with thousands more added daily. ORiGAMI is available as an open-science medical hypothesis generation tool - both as a web-service and an application programming interface (API) at http://hypothesis.ornl.gov . Since becoming an online service, ORIGAMI has enabled clinical subject-matter experts to: (i) discover the relationship between beta-blocker treatment and diabetic retinopathy; (ii) hypothesize that xylene is an environmental cancer-causing carcinogen and (iii) aid doctors with diagnosis of challenging cases when rare diseases manifest with common symptoms. In 2015, ORiGAMI was featured in the Historical Clinical Pathological Conference in Baltimore as a demonstration of artificial intelligence to medicine, IEEE/ACM Supercomputing and recognized as a Centennial Showcase Exhibit at the Radiological Society of North America (RSNA) Conference in Chicago. The final paper will describe the workflow built for the Cray Urika-XA and Urika-GD appliances that is able to reason with the knowledge of every published medical paper every time a clinical researcher uses the tool.« less

  14. Providing Access and Visualization to Global Cloud Properties from GEO Satellites

    NASA Astrophysics Data System (ADS)

    Chee, T.; Nguyen, L.; Minnis, P.; Spangenberg, D.; Palikonda, R.; Ayers, J. K.

    2015-12-01

    Providing public access to cloud macro and microphysical properties is a key concern for the NASA Langley Research Center Cloud and Radiation Group. This work describes a tool and method that allows end users to easily browse and access cloud information that is otherwise difficult to acquire and manipulate. The core of the tool is an application-programming interface that is made available to the public. One goal of the tool is to provide a demonstration to end users so that they can use the dynamically generated imagery as an input into their own work flows for both image generation and cloud product requisition. This project builds upon NASA Langley Cloud and Radiation Group's experience with making real-time and historical satellite cloud product imagery accessible and easily searchable. As we see the increasing use of virtual supply chains that provide additional value at each link there is value in making satellite derived cloud product information available through a simple access method as well as allowing users to browse and view that imagery as they need rather than in a manner most convenient for the data provider. Using the Open Geospatial Consortium's Web Processing Service as our access method, we describe a system that uses a hybrid local and cloud based parallel processing system that can return both satellite imagery and cloud product imagery as well as the binary data used to generate them in multiple formats. The images and cloud products are sourced from multiple satellites and also "merged" datasets created by temporally and spatially matching satellite sensors. Finally, the tool and API allow users to access information that spans the time ranges that our group has information available. In the case of satellite imagery, the temporal range can span the entire lifetime of the sensor.

  15. In Vitro and In Vivo Anti-tumoral Effects of the Flavonoid Apigenin in Malignant Mesothelioma

    PubMed Central

    Masuelli, Laura; Benvenuto, Monica; Mattera, Rosanna; Di Stefano, Enrica; Zago, Erika; Taffera, Gloria; Tresoldi, Ilaria; Giganti, Maria Gabriella; Frajese, Giovanni Vanni; Berardi, Ginevra; Modesti, Andrea; Bei, Roberto

    2017-01-01

    Malignant mesothelioma (MM) is a tumor arising from mesothelium. MM patients’ survival is poor. The polyphenol 4′,5,7,-trihydroxyflavone Apigenin (API) is a “multifunctional drug”. Several studies have demonstrated API anti-tumoral effects. However, little is known on the in vitro and in vivo anti-tumoral effects of API in MM. Thus, we analyzed the in vitro effects of API on cell proliferation, cell cycle regulation, pro-survival signaling pathways, apoptosis, and autophagy of human and mouse MM cells. We evaluated the in vivo anti-tumor activities of API in mice transplanted with MM #40a cells forming ascites. API inhibited in vitro MM cells survival, increased reactive oxygen species intracellular production and induced DNA damage. API activated apoptosis but not autophagy. API-induced apoptosis was sustained by the increase of Bax/Bcl-2 ratio, increase of p53 expression, activation of both caspase 9 and caspase 8, cleavage of PARP-1, and increase of the percentage of cells in subG1 phase. API treatment affected the phosphorylation of ERK1/2, JNK and p38 MAPKs in a cell-type specific manner, inhibited AKT phosphorylation, decreased c-Jun expression and phosphorylation, and inhibited NF-κB nuclear translocation. Intraperitoneal administration of API increased the median survival of C57BL/6 mice intraperitoneally transplanted with #40a cells and reduced the risk of tumor growth. Our findings may have important implications for the design of MM treatment using API. PMID:28674496

  16. Interactive Tools to Access the HELCATS Catalogues

    NASA Astrophysics Data System (ADS)

    Rouillard, Alexis; Plotnikov, Illya; Pinto, Rui; Génot, Vincent; Bouchemit, Myriam; Davies, Jackie

    2017-04-01

    The propagation tool is a web-based interface written in java that allows users to propagate Coronal Mass Ejections (CMEs), Corotating Interaction Regions (CIRs) and Solar Energetic Particles (SEPs) in the inner heliosphere. The tool displays unique datasets and catalogues through a 2-D visualisation of the trajectories of these heliospheric structures in relation to the orbital position of probes/planets and the pointing direction and extent of different imaging instruments. Summary plots of in-situ data or images of the solar corona and planetary aurorae stored at the CDPP, MEDOC and APIS databases, respectively, can be used to verify the presence of heliospheric structures at the estimated launch or impact times. A great novelty of the tool is the immediate visualisation of J-maps and the possibility to superpose on these maps the HELCATS CME and CIR catalogues.

  17. Interactive Tools to Access the HELCATS Catalogues

    NASA Astrophysics Data System (ADS)

    Rouillard, A.; Génot, V.; Bouchemit, M.; Pinto, R.

    2017-09-01

    The propagation tool is a web-based interface written in java that allows users to propagate Coronal Mass Ejections (CMEs), Corotating Interaction Regions (CIRs) and Solar Energetic Particles (SEPs) in the inner heliosphere. The tool displays unique datasets and catalogues through a 2-D visualisation of the trajectories of these heliospheric structures in relation to the orbital position of probes/planets and the pointing direction and extent of different imaging instruments. Summary plots of in-situ data or images of the solar corona and planetary aurorae stored at the CDPP, MEDOC and APIS databases, respectively, can be used to verify the presence of heliospheric structures at the estimated launch or impact times. A great novelty of the tool is the immediate visualisation of J-maps and the possibility to superpose on these maps the HELCATS CME and CIR catalogues.

  18. The putative serine protease inhibitor Api m 6 from Apis mellifera venom: recombinant and structural evaluation.

    PubMed

    Michel, Y; McIntyre, M; Ginglinger, H; Ollert, M; Cifuentes, L; Blank, S; Spillner, E

    2012-01-01

    Immunoglobulin (Ig) E-mediated reactions to honeybee venom can cause severe anaphylaxis, sometimes with fatal consequences. Detailed knowledge of the allergic potential of all venom components is necessary to ensure proper diagnosis and treatment of allergy and to gain a better understanding of the allergological mechanisms of insect venoms. Our objective was to undertake an immunochemical and structural evaluation of the putative low-molecular-weight serine protease inhibitor Api m 6, a component of honeybee venom. We recombinantly produced Api m 6 as a soluble protein in Escherichia coli and in Spodoptera frugiperda (Sf9) insect cells.We also assessed specific IgE reactivity of venom-sensitized patients with 2 prokaryotically produced Api m 6 variants using enzyme-linked immunosorbent assay. Moreover, we built a structural model ofApi m 6 and compared it with other protease inhibitor structures to gain insights into the function of Api m 6. In a population of 31 honeybee venom-allergic patients, 26% showed specific IgE reactivity with prokaryotically produced Api m 6, showing it to be a minor but relevant allergen. Molecular modeling of Api m 6 revealed a typical fold of canonical protease inhibitors, supporting the putative function of this venom allergen. Although Api m 6 has a highly variant surface charge, its epitope distribution appears to be similar to that of related proteins. Api m 6 is a honeybee venom component with IgE-sensitizing potential in a fraction of venom-allergic patients. Recombinant Api m 6 can help elucidate individual component-resolved reactivity profiles and increase our understanding of immune responses to low-molecular-weight allergens

  19. The tissue micro-array data exchange specification: a web based experience browsing imported data

    PubMed Central

    Nohle, David G; Hackman, Barbara A; Ayers, Leona W

    2005-01-01

    Background The AIDS and Cancer Specimen Resource (ACSR) is an HIV/AIDS tissue bank consortium sponsored by the National Cancer Institute (NCI) Division of Cancer Treatment and Diagnosis (DCTD). The ACSR offers to approved researchers HIV infected biologic samples and uninfected control tissues including tissue cores in micro-arrays (TMA) accompanied by de-identified clinical data. Researchers interested in the type and quality of TMA tissue cores and the associated clinical data need an efficient method for viewing available TMA materials. Because each of the tissue samples within a TMA has separate data including a core tissue digital image and clinical data, an organized, standard approach to producing, navigating and publishing such data is necessary. The Association for Pathology Informatics (API) extensible mark-up language (XML) TMA data exchange specification (TMA DES) proposed in April 2003 provides a common format for TMA data. Exporting TMA data into the proposed format offers an opportunity to implement the API TMA DES. Using our public BrowseTMA tool, we created a web site that organizes and cross references TMA lists, digital "virtual slide" images, TMA DES export data, linked legends and clinical details for researchers. Microsoft Excel® and Microsoft Word® are used to convert tabular clinical data and produce an XML file in the TMA DES format. The BrowseTMA tool contains Extensible Stylesheet Language Transformation (XSLT) scripts that convert XML data into Hyper-Text Mark-up Language (HTML) web pages with hyperlinks automatically added to allow rapid navigation. Results Block lists, virtual slide images, legends, clinical details and exports have been placed on the ACSR web site for 14 blocks with 1623 cores of 2.0, 1.0 and 0.6 mm sizes. Our virtual microscope can be used to view and annotate these TMA images. Researchers can readily navigate from TMA block lists to TMA legends and to clinical details for a selected tissue core. Exports for 11 blocks with 3812 cores from three other institutions were processed with the BrowseTMA tool. Fifty common data elements (CDE) from the TMA DES were used and 42 more created for site-specific data. Researchers can download TMA clinical data in the TMA DES format. Conclusion Virtual TMAs with clinical data can be viewed on the Internet by interested researchers using the BrowseTMA tool. We have organized our approach to producing, sorting, navigating and publishing TMA information to facilitate such review. We have converted Excel TMA data into TMA DES XML, and imported it and TMA DES XML from another institution into BrowseTMA to produce web pages that allow us to browse through the merged data. We proposed enhancements to the TMA DES as a result of this experience. We implemented improvements to the API TMA DES as a result of using exported data from several institutions. A document type definition was written for the API TMA DES (that optionally includes proposed enhancements). Independent validators can be used to check exports against the DTD (with or without the proposed enhancements). Linking tissue core images to readily navigable clinical data greatly improves the value of the TMA. PMID:16086837

  20. Ecology, Behaviour and Control of Apis cerana with a Focus on Relevance to the Australian Incursion

    PubMed Central

    Koetz, Anna H.

    2013-01-01

    Apis cerana Fabricius is endemic to most of Asia, where it has been used for honey production and pollination services for thousands of years. Since the 1980s, A. cerana has been introduced to areas outside its natural range (namely New Guinea, the Solomon Islands, and Australia), which sparked fears that it may become a pest species that could compete with, and negatively affect, native Australian fauna and flora, as well as commercially kept A. mellifera and commercial crops. This literature review is a response to these concerns and reviews what is known about the ecology and behaviour of A. cerana. Differences between temperate and tropical strains of A. cerana are reviewed, as are A. cerana pollination, competition between A. cerana and A. mellifera, and the impact and control strategies of introduced A. cerana, with a particular focus on gaps of current knowledge. PMID:26462524

  1. Pollination Services Provided by Bees in Pumpkin Fields Supplemented with Either Apis mellifera or Bombus impatiens or Not Supplemented

    PubMed Central

    Petersen, Jessica D.; Reiners, Stephen; Nault, Brian A.

    2013-01-01

    Pollinators provide an important service in many crops. Managed honey bees (Apis mellifera L.) are used to supplement pollination services provided by wild bees with the assumption that they will enhance pollination, fruit set and crop yield beyond the levels provided by the wild bees. Recent declines in managed honey bee populations have stimulated interest in finding alternative managed pollinators to service crops. In the eastern U.S., managed hives of the native common eastern bumble bee (Bombus impatiens Cresson) may be an excellent choice. To examine this issue, a comprehensive 2-yr study was conducted to compare fruit yield and bee visits to flowers in pumpkin (Cucurbita pepo L.) fields that were either supplemented with A. mellifera hives, B. impatiens hives or were not supplemented. We compared pumpkin yield, A. mellifera flower visitation frequency and B. impatiens flower visitation frequency between treatments. Results indicated that supplementing pumpkin fields with either A. mellifera or B. impatiens hives did not increase their visitation to pumpkin flowers or fruit yield compared with those that were not supplemented. Next, the relationship between frequency of pumpkin flower visitation by the most prominent bee species (Peponapis pruinosa (Say), B. impatiens and A. mellifera) and fruit yield was determined across all pumpkin fields sampled. Fruit yield increased as the frequency of flower visits by A. mellifera and B. impatiens increased in 2011 and 2012, respectively. These results suggest that supplementation with managed bees may not improve pumpkin production and that A. mellifera and B. impatiens are important pollinators of pumpkin in our system. PMID:23894544

  2. Pollination services provided by bees in pumpkin fields supplemented with either Apis mellifera or Bombus impatiens or not supplemented.

    PubMed

    Petersen, Jessica D; Reiners, Stephen; Nault, Brian A

    2013-01-01

    Pollinators provide an important service in many crops. Managed honey bees (Apis mellifera L.) are used to supplement pollination services provided by wild bees with the assumption that they will enhance pollination, fruit set and crop yield beyond the levels provided by the wild bees. Recent declines in managed honey bee populations have stimulated interest in finding alternative managed pollinators to service crops. In the eastern U.S., managed hives of the native common eastern bumble bee (Bombus impatiens Cresson) may be an excellent choice. To examine this issue, a comprehensive 2-yr study was conducted to compare fruit yield and bee visits to flowers in pumpkin (Cucurbita pepo L.) fields that were either supplemented with A. mellifera hives, B. impatiens hives or were not supplemented. We compared pumpkin yield, A. mellifera flower visitation frequency and B. impatiens flower visitation frequency between treatments. Results indicated that supplementing pumpkin fields with either A. mellifera or B. impatiens hives did not increase their visitation to pumpkin flowers or fruit yield compared with those that were not supplemented. Next, the relationship between frequency of pumpkin flower visitation by the most prominent bee species (Peponapis pruinosa (Say), B. impatiens and A. mellifera) and fruit yield was determined across all pumpkin fields sampled. Fruit yield increased as the frequency of flower visits by A. mellifera and B. impatiens increased in 2011 and 2012, respectively. These results suggest that supplementation with managed bees may not improve pumpkin production and that A. mellifera and B. impatiens are important pollinators of pumpkin in our system.

  3. Earth Science Mobile App Development for Non-Programmers

    NASA Astrophysics Data System (ADS)

    Oostra, D.; Crecelius, S.; Lewis, P.; Chambers, L. H.

    2012-08-01

    A number of cloud based visual development tools have emerged that provide methods for developing mobile applications quickly and without previous programming experience. The MY NASA DATA (MND) team would like to begin a discussion on how we can best leverage current mobile app technologies and available Earth science datasets. The MY NASA DATA team is developing an approach based on two main ideas. The first is to teach our constituents how to create mobile applications that interact with NASA datasets; the second is to provide web services or Application Programming Interfaces (APIs) that create sources of data that educators, students and scientists can use in their own mobile app development. This framework allows data providers to foster mobile application development and interaction while not becoming a software clearing house. MY NASA DATA's research has included meetings with local data providers, educators, libraries and individuals. A high level of interest has been identified from initial discussions and interviews. This overt interest combined with the marked popularity of mobile applications in our societies has created a new channel for outreach and communications with and between the science and educational communities.

  4. Real-time complex event processing for cloud resources

    NASA Astrophysics Data System (ADS)

    Adam, M.; Cordeiro, C.; Field, L.; Giordano, D.; Magnoni, L.

    2017-10-01

    The ongoing integration of clouds into the WLCG raises the need for detailed health and performance monitoring of the virtual resources in order to prevent problems of degraded service and interruptions due to undetected failures. When working in scale, the existing monitoring diversity can lead to a metric overflow whereby the operators need to manually collect and correlate data from several monitoring tools and frameworks, resulting in tens of different metrics to be constantly interpreted and analyzed per virtual machine. In this paper we present an ESPER based standalone application which is able to process complex monitoring events coming from various sources and automatically interpret data in order to issue alarms upon the resources’ statuses, without interfering with the actual resources and data sources. We will describe how this application has been used with both commercial and non-commercial cloud activities, allowing the operators to quickly be alarmed and react to misbehaving VMs and LHC experiments’ workflows. We will present the pattern analysis mechanisms being used, as well as the surrounding Elastic and REST API interfaces where the alarms are collected and served to users.

  5. Tenofovir disoproxil fumarate in Asian or Pacific Islander chronic hepatitis B patients with high viral load (≥ 9 log10 copies/ml).

    PubMed

    Fung, Scott; Gordon, Stuart C; Krastev, Zahary; Horban, Andrzej; Petersen, Jörg; Sperl, Jan; Gane, Edward; Jacobson, Ira M; Yee, Leland J; Dinh, Phillip; Martins, Eduardo B; Flaherty, John F; Kitrinos, Kathryn M; Dusheiko, Geoffrey; Trinh, Huy; Flisiak, Robert; Rustgi, Vinod K; Buti, Maria; Marcellin, Patrick

    2015-02-01

    We evaluated the antiviral response of Asian or Pacific Islander (API) patients with chronic hepatitis B (CHB) who had baseline high viral load (HVL), defined as pre-treatment hepatitis B virus (HBV) DNA ≥9 log10 copies/ml, following up to 288 weeks of tenofovir disoproxil fumarate (TDF) treatment. A total of 205 HBeAg-negative and HBeAg-positive self-described API patients received 48 weeks of TDF 300 mg (HVL n = 18) or adefovir dipivoxil 10 mg (HVL n = 15) in a blinded fashion, followed by open-label TDF for an additional 240 weeks. The proportions of HVL vs. non-HVL patients with HBV DNA <400 copies/ml were compared. Mean declines in HBV DNA were evaluated in API vs. non-API patients. Throughout the first 72 weeks of treatment, a smaller proportion of HVL API patients reached HBV DNA <400 copies/ml than non-HVL API patients. However, after this timepoint similar proportions of HVL and non-HVL API patients achieved HBV DNA <400 copies/ml (100% vs. 97%, respectively), which was maintained through week 288, where 92% of HVL patients and 99% of non-HVL API patients on treatment had HBV DNA <400 copies/ml. During the 288 weeks of treatment, API patients had similar mean HBV DNA declines as non-API patients, regardless of whether patients were HVL or non-HVL. No API HVL patient had persistent viremia at week 288. No resistance was detected among HVL or non-HVL patients. API patients with HVL CHB achieve HBV DNA <400 copies/ml with long-term TDF treatment; however, achieving viral suppression may take longer for HVL patients relative to non-HVL API patients. © 2014 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  6. Spatial and temporal analysis of Air Pollution Index and its timescale-dependent relationship with meteorological factors in Guangzhou, China, 2001-2011.

    PubMed

    Li, Li; Qian, Jun; Ou, Chun-Quan; Zhou, Ying-Xue; Guo, Cui; Guo, Yuming

    2014-07-01

    There is an increasing interest in spatial and temporal variation of air pollution and its association with weather conditions. We presented the spatial and temporal variation of Air Pollution Index (API) and examined the associations between API and meteorological factors during 2001-2011 in Guangzhou, China. A Seasonal-Trend Decomposition Procedure Based on Loess (STL) was used to decompose API. Wavelet analyses were performed to examine the relationships between API and several meteorological factors. Air quality has improved since 2005. APIs were highly correlated among five monitoring stations, and there were substantial temporal variations. Timescale-dependent relationships were found between API and a variety of meteorological factors. Temperature, relative humidity, precipitation and wind speed were negatively correlated with API, while diurnal temperature range and atmospheric pressure were positively correlated with API in the annual cycle. Our findings should be taken into account when determining air quality forecasts and pollution control measures. Copyright © 2014 Elsevier Ltd. All rights reserved.

  7. Integration and Exposure of Large Scale Computational Resources Across the Earth System Grid Federation (ESGF)

    NASA Astrophysics Data System (ADS)

    Duffy, D.; Maxwell, T. P.; Doutriaux, C.; Williams, D. N.; Chaudhary, A.; Ames, S.

    2015-12-01

    As the size of remote sensing observations and model output data grows, the volume of the data has become overwhelming, even to many scientific experts. As societies are forced to better understand, mitigate, and adapt to climate changes, the combination of Earth observation data and global climate model projects is crucial to not only scientists but to policy makers, downstream applications, and even the public. Scientific progress on understanding climate is critically dependent on the availability of a reliable infrastructure that promotes data access, management, and provenance. The Earth System Grid Federation (ESGF) has created such an environment for the Intergovernmental Panel on Climate Change (IPCC). ESGF provides a federated global cyber infrastructure for data access and management of model outputs generated for the IPCC Assessment Reports (AR). The current generation of the ESGF federated grid allows consumers of the data to find and download data with limited capabilities for server-side processing. Since the amount of data for future AR is expected to grow dramatically, ESGF is working on integrating server-side analytics throughout the federation. The ESGF Compute Working Team (CWT) has created a Web Processing Service (WPS) Application Programming Interface (API) to enable access scalable computational resources. The API is the exposure point to high performance computing resources across the federation. Specifically, the API allows users to execute simple operations, such as maximum, minimum, average, and anomalies, on ESGF data without having to download the data. These operations are executed at the ESGF data node site with access to large amounts of parallel computing capabilities. This presentation will highlight the WPS API, its capabilities, provide implementation details, and discuss future developments.

  8. Efficiency of local Indonesia honey bees (Apis cerana L.) and stingless bee (Trigona iridipennis) on tomato (Lycopersicon esculentum Mill.) pollination.

    PubMed

    Putra, Ramadhani Eka; Kinasih, Ida

    2014-01-01

    Tomato (Lycopersicon esculentum Mill.) is considered as one of major agricultural commodity of Indonesia farming. However, monthly production is unstable due to lack of pollination services. Common pollinator agent of tomatoes is bumblebees which is unsuitable for tropical climate of Indonesia and the possibility of alteration of local wild plant interaction with their pollinator. Indonesia is rich with wild bees and some of the species already domesticated for years with prospect as pollinating agent for tomatoes. This research aimed to assess the efficiency of local honey bee (Apis cerana L.) and stingless bee (Trigona iridipennis), as pollinator of tomato. During this research, total visitation rate and total numbers of pollinated flowers by honey bee and stingless bee were compared between them with bagged flowers as control. Total fruit production, average weight and size also measured in order to correlated pollination efficiency with quantity and quality of fruit produced. Result of this research showed that A. cerana has slightly higher rate of visitation (p>0.05) and significantly shorter handling time (p < 0.05) than T. iridipennis due to their larger colony demand and low reward provide by tomato flowers. However, honey bee pollinated tomato flowers more efficient pollinator than stingless bee (80.3 and 70.2% efficiency, respectively; p < 0.05) even though the average weight and size of tomatoes were similar (p>0.05). Based on the results, it is concluded that the use of Apis cerana and Trigona spp., for pollinating tomatoes in tropical climates could be an alternative to the use of non-native Apis mellifera and bumblebees (Bombus spp.). However, more researches are needed to evaluate the cost/benefit on large-scale farming and greenhouse pollination using both bees against other bee species and pollination methods.

  9. Status of emerging standards for data definitions and transfer in the petroleum industry

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Winczewski, L.M.

    1991-03-01

    Leading-edge hardware and software to store, retrieve, process, analyze, visualize, and interpret geoscience and petroleum data are improving continuously. A babel of definitions and formats for common industry data items limits the overall effectiveness of these computer-aided exploration and production tools. Custom data conversion required to load applications causes delays and exposes data content to error and degradation. Emerging industry-wide standards for management of geoscience and petroleum-related data are poised to overcome long-standing internal barriers to the full exploitation of these high-tech hardware/software systems. Industry technical organizations, such as AAPG, SEG, and API, have been actively pursuing industry-wide standards formore » data transfer, data definitions, and data models. These standard-defining groups are non-fee and solicit active participation from the entire petroleum community. The status of the most active of these groups is presented here. Data transfer standards are being pursued within AAPG (AAPG-B Data Transfer Standard), API (DLIS, for log data) and SEG (SEG-DEF, for seismic data). Converging data definitions, models, and glossaries are coming from the Petroleum Industry Data Dictionary Group (PIDD) and from subcommittees of the AAPG Computer Applications Committee. The National Computer Graphics Association is promoting development of standards for transfer of geographically oriented data. The API Well-Number standard is undergoing revision.« less

  10. Low level drug product API form analysis - Avalide tablet NIR quantitative method development and robustness challenges.

    PubMed

    Pan, Duohai; Crull, George; Yin, Shawn; Grosso, John

    2014-02-01

    Avalide(@), a medication used for the treatment of hypertension, is a combination of Irbesartan, and Hydrochlorothiazide. Irbesartan, one of the active pharmaceutical ingredients (API) in Avalide products, exists in two neat crystalline forms: Form A and Form B. Irbesartan Form A is the API form used in a wet granulation for the preparation of Avalide tablets. The presence of the less soluble Irbesartan Form B in Avalide tablets may result in the slower dissolution. In this paper, we have presented our work on the method development, verification and challenges of quantitatively detecting, via NIR and ssNMR, very small amounts of Irbesartan Form B in Avalide tablets. As part of the NIR method development and qualification, limit of detection, linearity and accuracy were examined. In addition, a limited study of the robustness of the method was conducted and a bias in the level of Form B was correlated to the ambient humidity. ssNMR, a primary method for the determination of polymorphic composition, was successfully used as an orthogonal technique to verify the accuracy of the NIR method and added to the confidence in the NIR method. The speed and efficiency of the NIR method make it a suitable and convenient tool for routine analysis of Avalide tablets for Form B in a QC environment. Copyright © 2013 Elsevier B.V. All rights reserved.

  11. 75 FR 58014 - Pipeline Safety: Information Collection Activity; Request for Comments

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-09-23

    ... detection systems must comply with the standards set out in American Petroleum Institute (API) publication API 1130. API 1130 requires operators to record and retain certain information regarding the operation and testing of CPM systems. Compliance with API 1130, including its recordkeeping requirements...

  12. The NASA Tournament Laboratory (NTL): Improving Data Access at PDS while Spreading Joy and Engaging Students through 16 Micro-Contests

    NASA Technical Reports Server (NTRS)

    LaMora, Andy; Raugh, A.; Erickson, K.; Grayzeck, E. J.; Knopf, W.; Morgan, T. H.

    2012-01-01

    NASA PDS hosts terabytes of valuable data from hundreds of data sources and spans decades of research. Data is stored on flat-file systems regulated through careful meta dictionaries. PDS's data is available to the public through its website which supports data searches through drill-down navigation. While the system returns data quickly, result sets in response to identical input differ depending on the drill-down path a user follows. To correct this Issue, to allow custom searching, and to improve general accessibility, PDS sought to create a new data structure and API, and to use them to build applications that are a joy to use and showcase the value of the data to students, teachers and citizens. PDS engaged TopCoder and Harvard Business School through the NTL to pursue these objectives in a pilot effort. Scope was limited to Small Bodies Node data. NTL analyzed data, proposed a solution, and implemented it through a series of micro-contests. Contest focused on different segments of the problem; conceptualization, architectural design, implementation, testing, etc. To demonstrate the utility of the completed solution, NTL developed web-based and mobile applications that can compare targets, regardless of mission. To further explore the potential of the solution NTL hosted "Mash-up" challenges that integrated the API with other publically available assets, to produce consumer and teaching applications, including an Augmented Reality iPad tool. Two contests were also posted to middle and high school students via the NoNameSite.com platform, and as a result of these contests, PDS/SBN has initiated a Facebook program. These contests defined and implemented a data warehouse with the necessary migration tools to transform legacy data, produced a public web interface for the new search, developed a public API, and produced four mobile applications that we expect to appeal to users both within and, without the academic community.

  13. PDS and NASA Tournament Laboratory Project to Engage Citizen Scientists and to Provide New Access to Cassini Data

    NASA Astrophysics Data System (ADS)

    Odess, Jennifer; Gordon, Mitch; Showalter, Mark; LaMora, Andy; Del Villar, Ambi; Raugh, Anne; Erickson, Kristen; Galica, Carol; Grayzeck, Ed; Morgan, Thomas; Knopf, Bill

    2014-11-01

    Jennifer Odess (1), Mitch Gordon (2), Mark Showalter (2), Andy LaMora (1), Ambi Del Villar (1), Anne Raugh (3), Kristen Erickson (4), Carol Galica (4), Ed Grayzeck (5), T. Morgan (5), and Bill Knopf (4)1. Appirio Top Coder, Inc2. SETI Institute3. University of Maryland4. NASA Headquarters5. Goddard Space Flight CenterThe Planetary Data System (PDS), working with the NASA Tournament Lab (NTL) and TopCoder® , is using challenge-based competition to generate new applications that increase both access to planetary data and discoverability—allowing users to “mine” data, and thus, to make new discoveries from data already “on the ground”. The first challenge-based completion was an optimized database and API for comet data at the PDS Small Bodies Node (SBN) in 2012. Since start-up, the installation at SBN has been tweaked to provide access to the comet data holdings of the SBN, and has introduced new users and new developers to PDS data. A follow-on contest using Cassini images from the PDS Rings Discipline Node, was designed to challenge the competitors to create new, more transparent, agile tools for public access to NASA’s planetary data, where “public” includes citizen scientists and educators. The experience gained with the API at SBN was applied to establishing a second installation at the PDS Planetary Rings Node (Rings), to serve as the basis to develop similar access tools at Rings to make the growing archive of Cassini images available through the API. The Cassini-Rings project had as its goal to develop a crowd-sourcing project with eventual application across the PDS holdings. From the contest results, a preliminary algorithm can detect known satellites hidden in Saturn’s rings which should prove valuable to programmers. The contest approach is also of potential use to educators for exercises studying the solar system. The progress to date and results of this citizen-scientist project will be discussed.

  14. The NASA Tournament Laboratory (“NTL”): Improving Data Access at PDS while Spreading Joy and Engaging Students through 16 Micro-Contests

    NASA Astrophysics Data System (ADS)

    LaMora, Andy; Raugh, A.; Erickson, K.; Grayzeck, E. J.; Knopf, W.; Lydon, M.; Lakhani, K.; Crusan, J.; Morgan, T. H.

    2012-10-01

    NASA PDS hosts terabytes of valuable data from hundreds of data sources and spans decades of research. Data is stored on flat-file systems regulated through careful meta dictionaries. PDS’s data is available to the public through its website which supports data searches through drill-down navigation. While the system returns data quickly, result sets in response to identical input differ depending on the drill-down path a user follows. To correct this issue, to allow custom searching, and to improve general accessibility, PDS sought to create a new data structure and API, and to use them to build applications that are a joy to use and showcase the value of the data to students, teachers and citizens. PDS engaged TopCoder and Harvard Business School through the NTL to pursue these objectives in a pilot effort. Scope was limited to Small Bodies Node data. NTL analyzed data, proposed a solution, and implemented it through a series of micro-contests. Contest focused on different segments of the problem; conceptualization, architectural design, implementation, testing, etc. To demonstrate the utility of the completed solution, NTL developed web-based and mobile applications that can compare targets, regardless of mission. To further explore the potential of the solution NTL hosted “Mash-up” challenges that integrated the API with other publically available assets, to produce consumer and teaching applications, including an Augmented Reality iPad tool. Two contests were also posted to middle and high school students via the NoNameSite.com platform, and as a result of these contests, PDS/SBN has initiated a Facebook program. These contests defined and implemented a data warehouse with the necessary migration tools to transform legacy data, produced a public web interface for the new search, developed a public API, and produced four mobile applications that we expect to appeal to users both within and without the academic community.

  15. Smoothing Data Friction through building Service Oriented Data Platforms

    NASA Astrophysics Data System (ADS)

    Wyborn, L. A.; Richards, C. J.; Evans, B. J. K.; Wang, J.; Druken, K. A.

    2017-12-01

    Data Friction has been commonly defined as the costs in time, energy and attention required to simply collect, check, store, move, receive, and access data. On average, researchers spend a significant fraction of their time finding the data for their research project and then reformatting it so that it can be used by the software application of their choice. There is an increasing role for both data repositories and software to be modernised to help reduce data friction in ways that support the better use of the data. Many generic data repositories simply accept data in the format as supplied: the key check is that the data have sufficient metadata to enable discovery and download. Few generic repositories have both the expertise and infrastructure to support the multiple domain specific requirements that facilitate the increasing need for integration and reusability. In contrast, major science domain-focused repositories are increasingly able to implement and enforce community endorsed best practices and guidelines that ensure reusability and harmonization of data for use within the community by offering semi-automated QC workflows to improve quality of submitted data. The most advanced of these science repositories now operate as service-oriented data platforms that extend the use of data across domain silos and increasingly provide server-side programmatically-enabled access to data via network protocols and community standard APIs. To provide this, more rigorous QA/QC procedures are needed to validate data against standards and community software and tools. This ensures that the data can be accessed in expected ways and also demonstrates that the data works across different (non-domain specific) packages, tools and programming languages deployed by the various user communities. In Australia, the National Computational Infrastructure (NCI) has created such a service-oriented data platform which is demonstrating how this approach can reduce data friction, servicing both individual domains as well as facilitating cross-domain collaboration. The approach has required an increase in effort for the repository to provide the additional expertise, so as to enable a better capability and efficient system which ultimately saves time by the individual researcher.

  16. The Ruby UCSC API: accessing the UCSC genome database using Ruby.

    PubMed

    Mishima, Hiroyuki; Aerts, Jan; Katayama, Toshiaki; Bonnal, Raoul J P; Yoshiura, Koh-ichiro

    2012-09-21

    The University of California, Santa Cruz (UCSC) genome database is among the most used sources of genomic annotation in human and other organisms. The database offers an excellent web-based graphical user interface (the UCSC genome browser) and several means for programmatic queries. A simple application programming interface (API) in a scripting language aimed at the biologist was however not yet available. Here, we present the Ruby UCSC API, a library to access the UCSC genome database using Ruby. The API is designed as a BioRuby plug-in and built on the ActiveRecord 3 framework for the object-relational mapping, making writing SQL statements unnecessary. The current version of the API supports databases of all organisms in the UCSC genome database including human, mammals, vertebrates, deuterostomes, insects, nematodes, and yeast.The API uses the bin index-if available-when querying for genomic intervals. The API also supports genomic sequence queries using locally downloaded *.2bit files that are not stored in the official MySQL database. The API is implemented in pure Ruby and is therefore available in different environments and with different Ruby interpreters (including JRuby). Assisted by the straightforward object-oriented design of Ruby and ActiveRecord, the Ruby UCSC API will facilitate biologists to query the UCSC genome database programmatically. The API is available through the RubyGem system. Source code and documentation are available at https://github.com/misshie/bioruby-ucsc-api/ under the Ruby license. Feedback and help is provided via the website at http://rubyucscapi.userecho.com/.

  17. The Ruby UCSC API: accessing the UCSC genome database using Ruby

    PubMed Central

    2012-01-01

    Background The University of California, Santa Cruz (UCSC) genome database is among the most used sources of genomic annotation in human and other organisms. The database offers an excellent web-based graphical user interface (the UCSC genome browser) and several means for programmatic queries. A simple application programming interface (API) in a scripting language aimed at the biologist was however not yet available. Here, we present the Ruby UCSC API, a library to access the UCSC genome database using Ruby. Results The API is designed as a BioRuby plug-in and built on the ActiveRecord 3 framework for the object-relational mapping, making writing SQL statements unnecessary. The current version of the API supports databases of all organisms in the UCSC genome database including human, mammals, vertebrates, deuterostomes, insects, nematodes, and yeast. The API uses the bin index—if available—when querying for genomic intervals. The API also supports genomic sequence queries using locally downloaded *.2bit files that are not stored in the official MySQL database. The API is implemented in pure Ruby and is therefore available in different environments and with different Ruby interpreters (including JRuby). Conclusions Assisted by the straightforward object-oriented design of Ruby and ActiveRecord, the Ruby UCSC API will facilitate biologists to query the UCSC genome database programmatically. The API is available through the RubyGem system. Source code and documentation are available at https://github.com/misshie/bioruby-ucsc-api/ under the Ruby license. Feedback and help is provided via the website at http://rubyucscapi.userecho.com/. PMID:22994508

  18. Datasets2Tools, repository and search engine for bioinformatics datasets, tools and canned analyses

    PubMed Central

    Torre, Denis; Krawczuk, Patrycja; Jagodnik, Kathleen M.; Lachmann, Alexander; Wang, Zichen; Wang, Lily; Kuleshov, Maxim V.; Ma’ayan, Avi

    2018-01-01

    Biomedical data repositories such as the Gene Expression Omnibus (GEO) enable the search and discovery of relevant biomedical digital data objects. Similarly, resources such as OMICtools, index bioinformatics tools that can extract knowledge from these digital data objects. However, systematic access to pre-generated ‘canned’ analyses applied by bioinformatics tools to biomedical digital data objects is currently not available. Datasets2Tools is a repository indexing 31,473 canned bioinformatics analyses applied to 6,431 datasets. The Datasets2Tools repository also contains the indexing of 4,901 published bioinformatics software tools, and all the analyzed datasets. Datasets2Tools enables users to rapidly find datasets, tools, and canned analyses through an intuitive web interface, a Google Chrome extension, and an API. Furthermore, Datasets2Tools provides a platform for contributing canned analyses, datasets, and tools, as well as evaluating these digital objects according to their compliance with the findable, accessible, interoperable, and reusable (FAIR) principles. By incorporating community engagement, Datasets2Tools promotes sharing of digital resources to stimulate the extraction of knowledge from biomedical research data. Datasets2Tools is freely available from: http://amp.pharm.mssm.edu/datasets2tools. PMID:29485625

  19. Datasets2Tools, repository and search engine for bioinformatics datasets, tools and canned analyses.

    PubMed

    Torre, Denis; Krawczuk, Patrycja; Jagodnik, Kathleen M; Lachmann, Alexander; Wang, Zichen; Wang, Lily; Kuleshov, Maxim V; Ma'ayan, Avi

    2018-02-27

    Biomedical data repositories such as the Gene Expression Omnibus (GEO) enable the search and discovery of relevant biomedical digital data objects. Similarly, resources such as OMICtools, index bioinformatics tools that can extract knowledge from these digital data objects. However, systematic access to pre-generated 'canned' analyses applied by bioinformatics tools to biomedical digital data objects is currently not available. Datasets2Tools is a repository indexing 31,473 canned bioinformatics analyses applied to 6,431 datasets. The Datasets2Tools repository also contains the indexing of 4,901 published bioinformatics software tools, and all the analyzed datasets. Datasets2Tools enables users to rapidly find datasets, tools, and canned analyses through an intuitive web interface, a Google Chrome extension, and an API. Furthermore, Datasets2Tools provides a platform for contributing canned analyses, datasets, and tools, as well as evaluating these digital objects according to their compliance with the findable, accessible, interoperable, and reusable (FAIR) principles. By incorporating community engagement, Datasets2Tools promotes sharing of digital resources to stimulate the extraction of knowledge from biomedical research data. Datasets2Tools is freely available from: http://amp.pharm.mssm.edu/datasets2tools.

  20. Detecting Runtime Anomalies in AJAX Applications through Trace Analysis

    DTIC Science & Technology

    2011-08-10

    statements by adding the instrumentation to the GWT UI classes, leaving the user code untouched. Some content management frameworks such as Drupal [12...Google web toolkit.” http://code.google.com/webtoolkit/. [12] “Form generation – drupal api.” http://api.drupal.org/api/group/form_api/6. 9

  1. Early Student Support for Application of Advanced Multi-Core Processor Technologies to Oceanographic Research

    DTIC Science & Technology

    2016-05-07

    REPORT DOCUMENTATION PAGE I . ... ... .. . ,...,.., ............. OMB No. 0704-0188 The public reporting burden for this collection of...Student Support for Appl ication of Advanced Multi- Core Processor N00014-12-1-0298 Technologies to Oceanographic Research Sb. GRANT NUMBER Sc...communications protocols (i.e. UART, I2C, and SPI), through the , ’ . handing off of the data to the server APis. By providing a common set of tools

  2. A RESTful API for Exchanging Materials Data in the AFLOWLIB.org Consortium

    DTIC Science & Technology

    2014-03-12

    of North Texas, Denton TX 4Materials Science, Electrical Engineering, Physics and Chemistry, Duke University, Durham NC, 27708 †On leave from the...software tools, input and output data are maintained remotely, lowering cost, improving ecological sustainability (saving electricity ) and increas- ing...enthalpy_formation_atom) – Description. Returns the formation enthalpy ∆HF per unit cell (∆HF atomic per atom). For compounds ANABNB · · · with NA + NB

  3. Distributed cyberinfrastructure tools for automated data processing of structural monitoring data

    NASA Astrophysics Data System (ADS)

    Zhang, Yilan; Kurata, Masahiro; Lynch, Jerome P.; van der Linden, Gwendolyn; Sederat, Hassan; Prakash, Atul

    2012-04-01

    The emergence of cost-effective sensing technologies has now enabled the use of dense arrays of sensors to monitor the behavior and condition of large-scale bridges. The continuous operation of dense networks of sensors presents a number of new challenges including how to manage such massive amounts of data that can be created by the system. This paper reports on the progress of the creation of cyberinfrastructure tools which hierarchically control networks of wireless sensors deployed in a long-span bridge. The internet-enabled cyberinfrastructure is centrally managed by a powerful database which controls the flow of data in the entire monitoring system architecture. A client-server model built upon the database provides both data-provider and system end-users with secured access to various levels of information of a bridge. In the system, information on bridge behavior (e.g., acceleration, strain, displacement) and environmental condition (e.g., wind speed, wind direction, temperature, humidity) are uploaded to the database from sensor networks installed in the bridge. Then, data interrogation services interface with the database via client APIs to autonomously process data. The current research effort focuses on an assessment of the scalability and long-term robustness of the proposed cyberinfrastructure framework that has been implemented along with a permanent wireless monitoring system on the New Carquinez (Alfred Zampa Memorial) Suspension Bridge in Vallejo, CA. Many data interrogation tools are under development using sensor data and bridge metadata (e.g., geometric details, material properties, etc.) Sample data interrogation clients including those for the detection of faulty sensors, automated modal parameter extraction.

  4. Mobile Phone Application Development for the Classroom

    NASA Astrophysics Data System (ADS)

    Lewis, P.; Oostra, D.; Crecelius, S.; Chambers, L. H.

    2012-08-01

    With smartphone sales currently surpassing laptop sales, it is hard not to think that these devices will have a place in the classroom. More specifically, with little to no monetary investment, classroom-centric mobile applications have the ability to suit the needs of teachers. Previously, programming such an item was a daunting task to the classroom teacher. But now, through the use of online visual tools, anyone has the ability to generate a mobile application to suit individual classroom needs. The "MY NASA DATA" (MND) project has begun work on such an application. Using online tools that are directed at the non-programmer, the team has developed two usable mobile applications ("apps") that fit right into the science classroom. The two apps generated include a cloud dichotomous key for cloud identification in the field, and an atmospheric science glossary to help with standardized testing key vocabulary and classroom assignments. Through the use of free online tools, teachers and students now have the ability to customize mobile applications to meet their individual needs. As an extension of the mobile applications, the MND team is planning web-based application programming interfaces (API's) that will be generated from data that is currently included in the MND Live Access Server. This will allow teachers and students to choose data sets that they want to include in the mobile application without having to populate the API themselves. Through the use of easy to understand online mobile app tutorials and MND data sets, teachers will have the ability to generate unit-specific mobile applications to further engage and empower students in the science classroom.

  5. Diagnosis and treatment of limb fractures associated with acute peripheral ischemia.

    PubMed

    Popescu, G I; Lupescu, O; Nagea, M; Patru, C

    2013-01-01

    Acute Peripheral Ischemia (API) is the most severe acute complication after both open and closed fractures, as ischemia compromises not only the vitality of the affected limb, but also the patient's life, because metabolic anaerobic changes following ischemia have serious local and general consequences. These explain why early diagnosis of API is very important for the prognosis of the traumatized limb.The authors analyse cases when API was not diagnosed immediately after trauma, but some time after the first examination, due to either low systolic BP or to late onset of API. The patients were analysed concerning the type of the fracture, the reason for delayed diagnosis of API, the moment of API diagnosis and the arterial injury. In all those cases, surgery was performed immediately after API diagnosis, in order to identify and treat the complex injuries(bone and vascular). Celsius.

  6. Infectivity and virulence of Nosema ceranae and Nosema apis in commercially available North American honey bees

    USDA-ARS?s Scientific Manuscript database

    Nosema ceranae infection is ubiquitous in western honey bees, Apis mellifera. In the US, displacement of N. apis in bee colonies suggests that N. ceranae has competitive advantages. Our study, however, showed that N. ceranae was significantly less infective and less virulent than N. apis in commerc...

  7. 47 CFR 61.46 - Adjustments to the API.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 3 2014-10-01 2014-10-01 false Adjustments to the API. 61.46 Section 61.46... Rules for Dominant Carriers § 61.46 Adjustments to the API. (a) Except as provided in paragraphs (d) and... cap local exchange carrier must calculate an API for each affected basket pursuant to the following...

  8. 47 CFR 61.46 - Adjustments to the API.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 3 2012-10-01 2012-10-01 false Adjustments to the API. 61.46 Section 61.46... Rules for Dominant Carriers § 61.46 Adjustments to the API. (a) Except as provided in paragraphs (d) and... cap local exchange carrier must calculate an API for each affected basket pursuant to the following...

  9. 47 CFR 61.46 - Adjustments to the API.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 47 Telecommunication 3 2011-10-01 2011-10-01 false Adjustments to the API. 61.46 Section 61.46... Rules for Dominant Carriers § 61.46 Adjustments to the API. (a) Except as provided in paragraphs (d) and... cap local exchange carrier must calculate an API for each affected basket pursuant to the following...

  10. 47 CFR 61.46 - Adjustments to the API.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 3 2010-10-01 2010-10-01 false Adjustments to the API. 61.46 Section 61.46... Rules for Dominant Carriers § 61.46 Adjustments to the API. (a) Except as provided in paragraphs (d) and... carrier must calculate an API for each affected basket pursuant to the following methodology: APIt = APIt...

  11. 47 CFR 61.46 - Adjustments to the API.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 3 2013-10-01 2013-10-01 false Adjustments to the API. 61.46 Section 61.46... Rules for Dominant Carriers § 61.46 Adjustments to the API. (a) Except as provided in paragraphs (d) and... cap local exchange carrier must calculate an API for each affected basket pursuant to the following...

  12. Solar Eclipse Computer API: Planning Ahead for August 2017

    NASA Astrophysics Data System (ADS)

    Bartlett, Jennifer L.; Chizek Frouard, Malynda; Lesniak, Michael V.; Bell, Steve

    2016-01-01

    With the total solar eclipse of 2017 August 21 over the continental United States approaching, the U.S. Naval Observatory (USNO) on-line Solar Eclipse Computer can now be accessed via an application programming interface (API). This flexible interface returns local circumstances for any solar eclipse in JavaScript Object Notation (JSON) that can be incorporated into third-party Web sites or applications. For a given year, it can also return a list of solar eclipses that can be used to build a more specific request for local circumstances. Over the course of a particular eclipse as viewed from a specific site, several events may be visible: the beginning and ending of the eclipse (first and fourth contacts), the beginning and ending of totality (second and third contacts), the moment of maximum eclipse, sunrise, or sunset. For each of these events, the USNO Solar Eclipse Computer reports the time, Sun's altitude and azimuth, and the event's position and vertex angles. The computer also reports the duration of the total phase, the duration of the eclipse, the magnitude of the eclipse, and the percent of the Sun obscured for a particular eclipse site. On-line documentation for using the API-enabled Solar Eclipse Computer, including sample calls, is available (http://aa.usno.navy.mil/data/docs/api.php). The same Web page also describes how to reach the Complete Sun and Moon Data for One Day, Phases of the Moon, Day and Night Across the Earth, and Apparent Disk of a Solar System Object services using API calls.For those who prefer using a traditional data input form, local circumstances can still be requested that way at http://aa.usno.navy.mil/data/docs/SolarEclipses.php. In addition, the 2017 August 21 Solar Eclipse Resource page (http://aa.usno.navy.mil/data/docs/Eclipse2017.php) consolidates all of the USNO resources for this event, including a Google Map view of the eclipse track designed by Her Majesty's Nautical Almanac Office (HMNAO). Looking further ahead, a 2024 April 8 Solar Eclipse Resource page (http://aa.usno.navy.mil/data/docs/Eclipse2024.php) is also available.

  13. Compositional mining of multiple object API protocols through state abstraction.

    PubMed

    Dai, Ziying; Mao, Xiaoguang; Lei, Yan; Qi, Yuhua; Wang, Rui; Gu, Bin

    2013-01-01

    API protocols specify correct sequences of method invocations. Despite their usefulness, API protocols are often unavailable in practice because writing them is cumbersome and error prone. Multiple object API protocols are more expressive than single object API protocols. However, the huge number of objects of typical object-oriented programs poses a major challenge to the automatic mining of multiple object API protocols: besides maintaining scalability, it is important to capture various object interactions. Current approaches utilize various heuristics to focus on small sets of methods. In this paper, we present a general, scalable, multiple object API protocols mining approach that can capture all object interactions. Our approach uses abstract field values to label object states during the mining process. We first mine single object typestates as finite state automata whose transitions are annotated with states of interacting objects before and after the execution of the corresponding method and then construct multiple object API protocols by composing these annotated single object typestates. We implement our approach for Java and evaluate it through a series of experiments.

  14. Compositional Mining of Multiple Object API Protocols through State Abstraction

    PubMed Central

    Mao, Xiaoguang; Qi, Yuhua; Wang, Rui; Gu, Bin

    2013-01-01

    API protocols specify correct sequences of method invocations. Despite their usefulness, API protocols are often unavailable in practice because writing them is cumbersome and error prone. Multiple object API protocols are more expressive than single object API protocols. However, the huge number of objects of typical object-oriented programs poses a major challenge to the automatic mining of multiple object API protocols: besides maintaining scalability, it is important to capture various object interactions. Current approaches utilize various heuristics to focus on small sets of methods. In this paper, we present a general, scalable, multiple object API protocols mining approach that can capture all object interactions. Our approach uses abstract field values to label object states during the mining process. We first mine single object typestates as finite state automata whose transitions are annotated with states of interacting objects before and after the execution of the corresponding method and then construct multiple object API protocols by composing these annotated single object typestates. We implement our approach for Java and evaluate it through a series of experiments. PMID:23844378

  15. A semi-automated workflow for biodiversity data retrieval, cleaning, and quality control

    PubMed Central

    Mathew, Cherian; Obst, Matthias; Vicario, Saverio; Haines, Robert; Williams, Alan R.; de Jong, Yde; Goble, Carole

    2014-01-01

    Abstract The compilation and cleaning of data needed for analyses and prediction of species distributions is a time consuming process requiring a solid understanding of data formats and service APIs provided by biodiversity informatics infrastructures. We designed and implemented a Taverna-based Data Refinement Workflow which integrates taxonomic data retrieval, data cleaning, and data selection into a consistent, standards-based, and effective system hiding the complexity of underlying service infrastructures. The workflow can be freely used both locally and through a web-portal which does not require additional software installations by users. PMID:25535486

  16. Physicochemical characterisation and investigation of the bonding mechanisms of API-titanate nanotube composites as new drug carrier systems.

    PubMed

    Sipos, Barbara; Pintye-Hódi, Klára; Kónya, Zoltán; Kelemen, András; Regdon, Géza; Sovány, Tamás

    2017-02-25

    Titanate nanotube (TNT) has recently been explored as a new carrier material for active pharmaceutical ingredients (API). The aim of the present work was to reveal the physicochemical properties of API-TNT composites, focusing on the interactions between the TNTs and the incorporated APIs. Drugs belonging to different Biopharmaceutical Classification System (BCS) classes were loaded into TNTs: diltiazem hydrochloride (BCS I.), diclofenac sodium (BCS II.), atenolol (BCS III.) and hydrochlorothiazide (BCS IV.). Experimental results demonstrated that it is feasible for spiral cross-sectioned titanate nanotubes to carry drugs and maintain their bioactivity. The structural properties of the composites were characterized by a range of analytical techniques, including FT-IR, DSC, TG-MS, etc. The interactions between APIs and TNTs were identified as electrostatic attractions, mainly dominated by hydrogen bonds. Based on the results, it can be stated that the strength of the association depends on the hydrogen donor strength of the API. The drug release of incorporated APIs was evaluated from compressed tablets and compared to that of pure APIs. Differences noticed in the dissolution profiles due to incorporation showed a correlation with the strength of interactions between the APIs and the TNTs observed in the above analytical studies. Copyright © 2016 Elsevier B.V. All rights reserved.

  17. OpenClimateGIS - A Web Service Providing Climate Model Data in Commonly Used Geospatial Formats

    NASA Astrophysics Data System (ADS)

    Erickson, T. A.; Koziol, B. W.; Rood, R. B.

    2011-12-01

    The goal of the OpenClimateGIS project is to make climate model datasets readily available in commonly used, modern geospatial formats used by GIS software, browser-based mapping tools, and virtual globes.The climate modeling community typically stores climate data in multidimensional gridded formats capable of efficiently storing large volumes of data (such as netCDF, grib) while the geospatial community typically uses flexible vector and raster formats that are capable of storing small volumes of data (relative to the multidimensional gridded formats). OpenClimateGIS seeks to address this difference in data formats by clipping climate data to user-specified vector geometries (i.e. areas of interest) and translating the gridded data on-the-fly into multiple vector formats. The OpenClimateGIS system does not store climate data archives locally, but rather works in conjunction with external climate archives that expose climate data via the OPeNDAP protocol. OpenClimateGIS provides a RESTful API web service for accessing climate data resources via HTTP, allowing a wide range of applications to access the climate data.The OpenClimateGIS system has been developed using open source development practices and the source code is publicly available. The project integrates libraries from several other open source projects (including Django, PostGIS, numpy, Shapely, and netcdf4-python).OpenClimateGIS development is supported by a grant from NOAA's Climate Program Office.

  18. Resolution and sensitivity of the eyes of the Asian honeybees Apis florea, Apis cerana and Apis dorsata.

    PubMed

    Somanathan, Hema; Warrant, Eric J; Borges, Renee M; Wallén, Rita; Kelber, Almut

    2009-08-01

    Bees of the genus Apis are important foragers of nectar and pollen resources. Although the European honeybee, Apis mellifera, has been well studied with respect to its sensory abilities, learning behaviour and role as pollinators, much less is known about the other Apis species. We studied the anatomical spatial resolution and absolute sensitivity of the eyes of three sympatric species of Asian honeybees, Apis cerana, Apis florea and Apis dorsata and compared them with the eyes of A. mellifera. Of these four species, the giant honeybee A. dorsata (which forages during moonlit nights) has the lowest spatial resolution and the most sensitive eyes, followed by A. mellifera, A. cerana and the dwarf honeybee, A. florea (which has the smallest acceptance angles and the least sensitive eyes). Moreover, unlike the strictly diurnal A. cerana and A. florea, A. dorsata possess large ocelli, a feature that it shares with all dim-light bees. However, the eyes of the facultatively nocturnal A. dorsata are much less sensitive than those of known obligately nocturnal bees such as Megalopta genalis in Panama and Xylocopa tranquebarica in India. The differences in sensitivity between the eyes of A. dorsata and other strictly diurnal Apis species cannot alone explain why the former is able to fly, orient and forage at half-moon light levels. We assume that additional neuronal adaptations, as has been proposed for A. mellifera, M. genalis and X. tranquebarica, might exist in A. dorsata.

  19. Improving the API dissolution rate during pharmaceutical hot-melt extrusion I: Effect of the API particle size, and the co-rotating, twin-screw extruder screw configuration on the API dissolution rate.

    PubMed

    Li, Meng; Gogos, Costas G; Ioannidis, Nicolas

    2015-01-15

    The dissolution rate of the active pharmaceutical ingredients in pharmaceutical hot-melt extrusion is the most critical elementary step during the extrusion of amorphous solid solutions - total dissolution has to be achieved within the short residence time in the extruder. Dissolution and dissolution rates are affected by process, material and equipment variables. In this work, we examine the effect of one of the material variables and one of the equipment variables, namely, the API particle size and extruder screw configuration on the API dissolution rate, in a co-rotating, twin-screw extruder. By rapidly removing the extruder screws from the barrel after achieving a steady state, we collected samples along the length of the extruder screws that were characterized by polarized optical microscopy (POM) and differential scanning calorimetry (DSC) to determine the amount of undissolved API. Analyses of samples indicate that reduction of particle size of the API and appropriate selection of screw design can markedly improve the dissolution rate of the API during extrusion. In addition, angle of repose measurements and light microscopy images show that the reduction of particle size of the API can improve the flowability of the physical mixture feed and the adhesiveness between its components, respectively, through dry coating of the polymer particles by the API particles. Copyright © 2014. Published by Elsevier B.V.

  20. Comparison of Vitek Matrix-assisted Laser Desorption/Ionization Time-of-Flight Mass Spectrometry Versus Conventional Methods in Candida Identification.

    PubMed

    Keçeli, Sema Aşkın; Dündar, Devrim; Tamer, Gülden Sönmez

    2016-02-01

    Candida species are generally identified by conventional methods such as germ tube or morphological appearance on corn meal agar, biochemical methods using API kits and molecular biological methods. Alternative to these methods, rapid and accurate identification methods of microorganisms called matrix-assisted laser desorption/ionization time-of-flight mass spectrometry (MALDİ-TOF MS) has recently been described. In this study, Candida identification results by API Candida kit, API 20C AUX kit and identifications on corn meal agar (CMA) are compared with the results obtained on Vitek-MS. All results were confirmed by sequencing internal transcribed spacer (ITS) regions of rDNA. Totally, 97 Candida strains were identified by germ tube test, CMA, API and Vitek-MS. Vitek-MS results were compatible with 74.2 % of API 20C AUX and 81.4 % of CMA results. The difference between the results of API Candida and API 20C AUX was detected. The ratio of discrepancy between Vitek-MS and API 20C AUX was 25.8 %. Candida species mostly identified as C. famata or C. tropicalis by and not compatible with API kits were identified as C. albicans by Vitek-MS. Sixteen Candida species having discrepant results with Vitek-MS, API or CMA were randomly chosen, and ITS sequence analysis was performed. The results of sequencing were compatible 56.2 % with API 20C AUX, 50 % with CMA and 93.7 % with Vitek-MS. When compared with conventional identification methods, MS results are more reliable and rapid for Candida identification. MS system may be used as routine identification method in clinical microbiology laboratories.

Top